Files
hermes-sync/skills/mlops/training/axolotl/references/api.md

118 KiB
Raw Blame History

Axolotl - Api

Pages: 150


cli.cloud.modal_

URL: https://docs.axolotl.ai/docs/api/cli.cloud.modal_.html

Contents:

  • cli.cloud.modal_
  • Classes
    • ModalCloud
  • Functions
    • run_cmd

Modal Cloud support from CLI

Modal Cloud implementation.

Run a command inside a folder, with Modal Volume reloading before and commit on success.

Examples:

Example 1 (python):

cli.cloud.modal_.ModalCloud(config, app=None)

Example 2 (python):

cli.cloud.modal_.run_cmd(cmd, run_folder, volumes=None)

core.trainers.base

URL: https://docs.axolotl.ai/docs/api/core.trainers.base.html

Contents:

  • core.trainers.base
  • Classes
    • AxolotlTrainer
      • Methods
        • log
          • Parameters
        • push_to_hub
        • store_metrics
          • Parameters

Module for customized trainers

Extend the base Trainer for axolotl helpers

Log logs on the various objects watching training, including stored metrics.

Overwrite the push_to_hub method in order to force-add the tags when pushing the model on the Hub. Please refer to ~transformers.Trainer.push_to_hub for more details.

Store metrics with specified reduction type.

Examples:

Example 1 (python):

core.trainers.base.AxolotlTrainer(
    *_args,
    bench_data_collator=None,
    eval_data_collator=None,
    dataset_tags=None,
    **kwargs,
)

Example 2 (python):

core.trainers.base.AxolotlTrainer.log(logs, start_time=None)

Example 3 (python):

core.trainers.base.AxolotlTrainer.push_to_hub(*args, **kwargs)

Example 4 (python):

core.trainers.base.AxolotlTrainer.store_metrics(
    metrics,
    train_eval='train',
    reduction='mean',
)

prompt_strategies.input_output

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.input_output.html

Contents:

  • prompt_strategies.input_output
  • Classes
    • RawInputOutputPrompter
    • RawInputOutputStrategy

prompt_strategies.input_output

Module for plain input/output prompt pairs

prompter for raw i/o data

Prompt Strategy class for input/output pairs

Examples:

Example 1 (python):

prompt_strategies.input_output.RawInputOutputPrompter()

Example 2 (python):

prompt_strategies.input_output.RawInputOutputStrategy(
    *args,
    eos_token=None,
    **kwargs,
)

prompt_strategies.completion

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.completion.html

Contents:

  • prompt_strategies.completion
  • Classes
    • CompletionPromptTokenizingStrategy
    • CompletionPrompter

prompt_strategies.completion

Basic completion text

Tokenizing strategy for Completion prompts.

Prompter for completion

Examples:

Example 1 (python):

prompt_strategies.completion.CompletionPromptTokenizingStrategy(
    *args,
    max_length=None,
    **kwargs,
)

Example 2 (python):

prompt_strategies.completion.CompletionPrompter()

utils.collators.core

URL: https://docs.axolotl.ai/docs/api/utils.collators.core.html

Contents:

  • utils.collators.core

basic shared collator constants


monkeypatch.data.batch_dataset_fetcher

URL: https://docs.axolotl.ai/docs/api/monkeypatch.data.batch_dataset_fetcher.html

Contents:

  • monkeypatch.data.batch_dataset_fetcher
  • Functions
    • apply_multipack_dataloader_patch
    • patch_fetchers
    • patched_worker_loop
    • remove_multipack_dataloader_patch

monkeypatch.data.batch_dataset_fetcher

Monkey patches for the dataset fetcher to handle batches of packed indexes.

This patch allows DataLoader to correctly process batches that contain multiple bins of packed sequences.

Apply patches to PyTorchs DataLoader components.

Worker loop that ensures patches are applied in worker processes.

Remove the monkeypatch and restore original PyTorch DataLoader behavior.

Examples:

Example 1 (python):

monkeypatch.data.batch_dataset_fetcher.apply_multipack_dataloader_patch()

Example 2 (python):

monkeypatch.data.batch_dataset_fetcher.patch_fetchers()

Example 3 (python):

monkeypatch.data.batch_dataset_fetcher.patched_worker_loop(*args, **kwargs)

Example 4 (python):

monkeypatch.data.batch_dataset_fetcher.remove_multipack_dataloader_patch()

core.datasets.chat

URL: https://docs.axolotl.ai/docs/api/core.datasets.chat.html

Contents:

  • core.datasets.chat
  • Classes
    • TokenizedChatDataset

Tokenized chat dataset

Examples:

Example 1 (python):

core.datasets.chat.TokenizedChatDataset(
    data,
    model_transform,
    *args,
    message_transform=None,
    formatter=None,
    process_count=None,
    keep_in_memory=False,
    **kwargs,
)

utils.freeze

URL: https://docs.axolotl.ai/docs/api/utils.freeze.html

Contents:

  • utils.freeze
  • Classes
    • LayerNamePattern
      • Methods
        • match
  • Functions
    • freeze_layers_except

module to freeze/unfreeze parameters by name

Represents a regex pattern for layer names, potentially including a parameter index range.

Checks if the given layer name matches the regex pattern.

Parameters: - name (str): The layer name to check.

Returns: - bool: True if the layer name matches the pattern, False otherwise.

Freezes all layers of the given model except for the layers that match given regex patterns. Periods in the patterns are treated as literal periods, not as wildcard characters.

Parameters: - model (nn.Module): The PyTorch model to be modified. - regex_patterns (list of str): List of regex patterns to match layer names to keep unfrozen. Note that you cannot use a dot as a wildcard character in the patterns since it is reserved for separating layer names. Also, to match the entire layer name, the pattern should start with “^” and end with “(", otherwise it will match any part of the layer name. The range pattern part is optional and it is not compiled as a regex pattern which means you must put ")” before the range pattern if you want to match the entire layer name. E.g., [“^model.embed_tokens.weight([:32000]", "layers.2[0-9]+.block_sparse_moe.gate.[a-z]+)”]

Returns: None; the model is modified in place.

Examples:

Example 1 (python):

utils.freeze.LayerNamePattern(pattern)

Example 2 (python):

utils.freeze.LayerNamePattern.match(name)

Example 3 (python):

utils.freeze.freeze_layers_except(model, regex_patterns)

monkeypatch.unsloth_

URL: https://docs.axolotl.ai/docs/api/monkeypatch.unsloth_.html

Contents:

  • monkeypatch.unsloth_

module for patching with unsloth optimizations


utils.schemas.datasets

URL: https://docs.axolotl.ai/docs/api/utils.schemas.datasets.html

Contents:

  • utils.schemas.datasets
  • Classes
    • DPODataset
    • KTODataset
    • PretrainingDataset
    • SFTDataset
      • Methods
        • handle_legacy_message_fields
    • StepwiseSupervisedDataset
    • UserDefinedDPOType

utils.schemas.datasets

Pydantic models for datasets-related configuration

DPO configuration subset

KTO configuration subset

Pretraining dataset configuration subset

SFT configuration subset

Handle backwards compatibility between legacy message field mapping and new property mapping system.

Stepwise supervised dataset configuration subset

User defined typing for DPO

User defined typing for KTO

Structure for user defined prompt types

Examples:

Example 1 (python):

utils.schemas.datasets.DPODataset()

Example 2 (python):

utils.schemas.datasets.KTODataset()

Example 3 (python):

utils.schemas.datasets.PretrainingDataset()

Example 4 (python):

utils.schemas.datasets.SFTDataset()

core.chat.format.llama3x

URL: https://docs.axolotl.ai/docs/api/core.chat.format.llama3x.html

Contents:

  • core.chat.format.llama3x

core.chat.format.llama3x

Llama 3.x chat formatting functions for MessageContents


datasets

URL: https://docs.axolotl.ai/docs/api/datasets.html

Contents:

  • datasets
  • Classes
    • TokenizedPromptDataset
      • Parameters

Module containing dataset functionality.

We want this to be a wrapper for an existing dataset that we have loaded. Lets use the concept of middlewares to wrap each dataset. Well use the collators later on to pad the datasets.

Dataset that returns tokenized prompts from a stream of text files.

Examples:

Example 1 (python):

datasets.TokenizedPromptDataset(
    prompt_tokenizer,
    dataset,
    process_count=None,
    keep_in_memory=False,
    **kwargs,
)

prompt_strategies.bradley_terry.llama3

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.bradley_terry.llama3.html

Contents:

  • prompt_strategies.bradley_terry.llama3
  • Functions
    • icr

prompt_strategies.bradley_terry.llama3

chatml transforms for datasets with system, input, chosen, rejected to match llama3 chat template

chatml transforms for datasets with system, input, chosen, rejected ex. https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs

Examples:

Example 1 (python):

prompt_strategies.bradley_terry.llama3.icr(cfg, **kwargs)

common.datasets

URL: https://docs.axolotl.ai/docs/api/common.datasets.html

Contents:

  • common.datasets
  • Classes
    • TrainDatasetMeta
  • Functions
    • load_datasets
      • Parameters
      • Returns
    • load_preference_datasets
      • Parameters
      • Returns

Dataset loading utilities.

Dataclass with fields for training and validation datasets and metadata.

Loads one or more training or evaluation datasets, calling axolotl.utils.data.prepare_datasets. Optionally, logs out debug information.

Loads one or more training or evaluation datasets for RL training using paired preference data, calling axolotl.utils.data.rl.prepare_preference_datasets. Optionally, logs out debug information.

Randomly sample num_samples samples with replacement from dataset.

Examples:

Example 1 (python):

common.datasets.TrainDatasetMeta(
    train_dataset,
    eval_dataset=None,
    total_num_steps=None,
)

Example 2 (python):

common.datasets.load_datasets(cfg, cli_args=None, debug=False)

Example 3 (python):

common.datasets.load_preference_datasets(cfg, cli_args=None)

Example 4 (python):

common.datasets.sample_dataset(dataset, num_samples)

cli.train

URL: https://docs.axolotl.ai/docs/api/cli.train.html

Contents:

  • cli.train
  • Functions
    • do_cli
      • Parameters
    • do_train
      • Parameters

CLI to run training on a model.

Parses axolotl config, CLI args, and calls do_train.

Trains a transformers model by first loading the dataset(s) specified in the axolotl config, and then calling axolotl.train.train. Also runs the plugin managers post_train_unload once training completes.

Examples:

Example 1 (python):

cli.train.do_cli(config=Path('examples/'), **kwargs)

Example 2 (python):

cli.train.do_train(cfg, cli_args)

cli.utils.fetch

URL: https://docs.axolotl.ai/docs/api/cli.utils.fetch.html

Contents:

  • cli.utils.fetch
  • Functions
    • fetch_from_github
      • Parameters

Utilities for axolotl fetch CLI command.

Sync files from a specific directory in the GitHub repository. Only downloads files that dont exist locally or have changed.

Examples:

Example 1 (python):

cli.utils.fetch.fetch_from_github(dir_prefix, dest_dir=None, max_workers=5)

utils.tokenization

URL: https://docs.axolotl.ai/docs/api/utils.tokenization.html

Contents:

  • utils.tokenization
  • Functions
    • color_token_for_rl_debug
    • process_tokens_for_rl_debug

Module for tokenization utilities

Helper function to color tokens based on their type.

Helper function to process and color tokens.

Examples:

Example 1 (python):

utils.tokenization.color_token_for_rl_debug(
    decoded_token,
    encoded_token,
    color,
    text_only,
)

Example 2 (python):

utils.tokenization.process_tokens_for_rl_debug(
    tokens,
    color,
    tokenizer,
    text_only,
)

core.trainers.grpo.sampler

URL: https://docs.axolotl.ai/docs/api/core.trainers.grpo.sampler.html

Contents:

  • core.trainers.grpo.sampler
  • Classes
    • SequenceParallelRepeatRandomSampler
      • Parameters
      • Methods
        • set_epoch
          • Parameters

core.trainers.grpo.sampler

Repeat random sampler (similar to the one implemented in https://github.com/huggingface/trl/blob/main/trl/trainer/grpo_trainer.py) that adds sequence parallelism functionality; i.e., duplicating data across ranks in the same sequence parallel group.

Sampler for GRPO training with sequence parallelism.

This sampler ensures: - Ranks in the same sequence parallel (SP) group receive identical data. - Each index is repeated multiple times for sampling different completions. - Entire batches are repeated for reuse in multiple updates. - Data is properly distributed across SP groups.

In the table below, the values represent dataset indices. Each SP group has context_parallel_size = 2 GPUs working together on the same data. There are 2 SP groups (SP0 and SP1), with world_size = 4 total GPUs.

grad_accum=2 ▲ ▲ 0 0 [0 0 0 1 1 1] [2 2 2 3 3 3] <- SP groups get different data ▼ | 0 1 [0 0 0 1 1 1] [2 2 2 3 3 3] <- Same data for each SP group GPU | | 1 2 [0 0 0 1 1 1] [2 2 2 3 3 3] <- Repeat same indices for iterations num_iterations=2 ▼ 1 3 [0 0 0 1 1 1] [2 2 2 3 3 3] <- When using gradient accumulation

Sets the epoch for this sampler.

Examples:

Example 1 (python):

core.trainers.grpo.sampler.SequenceParallelRepeatRandomSampler(
    dataset,
    mini_repeat_count,
    world_size,
    rank,
    batch_size=1,
    repeat_count=1,
    context_parallel_size=1,
    shuffle=True,
    seed=0,
    drop_last=False,
)

Example 2 (unknown):

Sequence Parallel Groups
                                |       SP0        |       SP1        |
                                |  GPU 0  |  GPU 1 |  GPU 2  |  GPU 3 |
            global_step  step    <---> mini_repeat_count=3
                                    <----------> batch_size=2 per SP group

Example 3 (unknown):

2       4         [4 4 4  5 5 5]     [6 6 6  7 7 7]   <- New batch of data indices
                 2       5         [4 4 4  5 5 5]     [6 6 6  7 7 7]
                                    ...

Example 4 (python):

core.trainers.grpo.sampler.SequenceParallelRepeatRandomSampler.set_epoch(epoch)

evaluate

URL: https://docs.axolotl.ai/docs/api/evaluate.html

Contents:

  • evaluate
  • Functions
    • evaluate
      • Parameters
      • Returns
    • evaluate_dataset
      • Parameters
      • Returns

Module for evaluating models.

Evaluate a model on training and validation datasets.

Helper function to evaluate a single dataset.

Examples:

Example 1 (python):

evaluate.evaluate(cfg, dataset_meta)

Example 2 (python):

evaluate.evaluate_dataset(trainer, dataset, dataset_type, flash_optimum=False)

utils.optimizers.adopt

URL: https://docs.axolotl.ai/docs/api/utils.optimizers.adopt.html

Contents:

  • utils.optimizers.adopt
  • Functions
    • adopt

utils.optimizers.adopt

Copied from https://github.com/iShohei220/adopt

ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate (2024) Taniguchi, Shohei and Harada, Keno and Minegishi, Gouki and Oshima, Yuta and Jeong, Seong Cheol and Nagahara, Go and Iiyama, Tomoshi and Suzuki, Masahiro and Iwasawa, Yusuke and Matsuo, Yutaka

Functional API that performs ADOPT algorithm computation.

Examples:

Example 1 (python):

utils.optimizers.adopt.adopt(
    params,
    grads,
    exp_avgs,
    exp_avg_sqs,
    state_steps,
    foreach=None,
    capturable=False,
    differentiable=False,
    fused=None,
    grad_scale=None,
    found_inf=None,
    has_complex=False,
    *,
    beta1,
    beta2,
    lr,
    clip_lambda,
    weight_decay,
    decouple,
    eps,
    maximize,
)

prompt_tokenizers

URL: https://docs.axolotl.ai/docs/api/prompt_tokenizers.html

Contents:

  • prompt_tokenizers
  • Classes
    • AlpacaMultipleChoicePromptTokenizingStrategy
    • AlpacaPromptTokenizingStrategy
    • AlpacaReflectionPTStrategy
    • DatasetWrappingStrategy
    • GPTeacherPromptTokenizingStrategy
    • InstructionPromptTokenizingStrategy
    • InvalidDataException
    • JeopardyPromptTokenizingStrategy

Module containing PromptTokenizingStrategy and Prompter classes

Tokenizing strategy for Alpaca Multiple Choice prompts.

Tokenizing strategy for Alpaca prompts.

Tokenizing strategy for Alpaca Reflection prompts.

Abstract class for wrapping datasets for Chat Messages

Tokenizing strategy for GPTeacher prompts.

Tokenizing strategy for instruction-based prompts.

Exception raised when the data is invalid

Tokenizing strategy for Jeopardy prompts.

Tokenizing strategy for NomicGPT4All prompts.

Tokenizing strategy for OpenAssistant prompts.

Abstract class for tokenizing strategies

Tokenizing strategy for Reflection prompts.

Tokenizing strategy for SummarizeTLDR prompts.

Parses the tokenized prompt and append the tokenized input_ids, attention_mask and labels to the result

Returns the default values for the tokenize prompt function

Examples:

Example 1 (python):

prompt_tokenizers.AlpacaMultipleChoicePromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 2 (python):

prompt_tokenizers.AlpacaPromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 3 (python):

prompt_tokenizers.AlpacaReflectionPTStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 4 (python):

prompt_tokenizers.DatasetWrappingStrategy()

cli.art

URL: https://docs.axolotl.ai/docs/api/cli.art.html

Contents:

  • cli.art
  • Functions
    • print_axolotl_text_art

Axolotl ASCII logo utils.

Prints axolotl ASCII art.

Examples:

Example 1 (python):

cli.art.print_axolotl_text_art()

utils.callbacks.perplexity

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.perplexity.html

Contents:

  • utils.callbacks.perplexity
  • Classes
    • Perplexity
      • Methods
        • compute

utils.callbacks.perplexity

callback to calculate perplexity as an evaluation metric.

Calculate perplexity as defined in https://huggingface.co/docs/transformers/en/perplexity. This is a custom variant that doesnt re-tokenize the input or re-load the model.

Compute perplexity in a fixed length sliding window across the sequence.

Examples:

Example 1 (python):

utils.callbacks.perplexity.Perplexity(tokenizer, max_seq_len, stride=512)

Example 2 (python):

utils.callbacks.perplexity.Perplexity.compute(model, references=None)

cli.utils.train

URL: https://docs.axolotl.ai/docs/api/cli.utils.train.html

Contents:

  • cli.utils.train
  • Functions
    • build_command
      • Parameters
      • Returns
    • generate_config_files
      • Parameters
    • launch_training

Utilities for axolotl train CLI command.

Build command list from base command and options.

Generate list of configuration files to process. Yields a tuple of the configuration file name and a boolean indicating whether this is a group of configurations (i.e., a sweep).

Execute training with the given configuration.

Examples:

Example 1 (python):

cli.utils.train.build_command(base_cmd, options)

Example 2 (python):

cli.utils.train.generate_config_files(config, sweep)

Example 3 (python):

cli.utils.train.launch_training(
    cfg_file,
    launcher,
    cloud,
    kwargs,
    launcher_args=None,
    use_exec=False,
)

cli.vllm_serve

URL: https://docs.axolotl.ai/docs/api/cli.vllm_serve.html

Contents:

  • cli.vllm_serve
  • Classes
    • AxolotlScriptArguments
  • Functions
    • do_vllm_serve
      • Returns

CLI to start the vllm server for online RL

Additional arguments for the VLLM server

Starts the VLLM server for serving LLM models used for online RL

Args :param cfg: Parsed doct of the YAML config :param cli_args: dict of additional command-line arguments of type VllmServeCliArgs

Examples:

Example 1 (python):

cli.vllm_serve.AxolotlScriptArguments(
    reasoning_parser='',
    enable_reasoning=None,
)

Example 2 (python):

cli.vllm_serve.do_vllm_serve(config, cli_args)

convert

URL: https://docs.axolotl.ai/docs/api/convert.html

Contents:

  • convert
  • Classes
    • FileReader
    • FileWriter
    • JsonParser
    • JsonToJsonlConverter
    • JsonlSerializer
    • StdoutWriter

Module containing File Reader, File Writer, Json Parser, and Jsonl Serializer classes

Reads a file and returns its contents as a string

Writes a string to a file

Parses a string as JSON and returns the result

Converts a JSON file to JSONL

Serializes a list of JSON objects into a JSONL string

Writes a string to stdout

Examples:

Example 1 (python):

convert.FileReader()

Example 2 (python):

convert.FileWriter(file_path)

Example 3 (python):

convert.JsonParser()

Example 4 (python):

convert.JsonToJsonlConverter(
    file_reader,
    file_writer,
    json_parser,
    jsonl_serializer,
)

monkeypatch.utils

URL: https://docs.axolotl.ai/docs/api/monkeypatch.utils.html

Contents:

  • monkeypatch.utils
  • Functions
    • get_cu_seqlens
    • get_cu_seqlens_from_pos_ids
    • mask_2d_to_4d

Shared utils for the monkeypatches

generate a cumulative sequence length mask for flash attention using attn mask

generate a cumulative sequence length mask for flash attention using pos ids

Expands attention_mask from [bsz, seq_len] to [bsz, 1, tgt_seq_len, src_seq_len]. This expansion handles packed sequences so that sequences share the same attention mask integer value when they attend to each other within that sequence. This expansion transforms the mask to lower triangular form to prevent future peeking.

Examples:

Example 1 (python):

monkeypatch.utils.get_cu_seqlens(attn_mask)

Example 2 (python):

monkeypatch.utils.get_cu_seqlens_from_pos_ids(position_ids)

Example 3 (python):

monkeypatch.utils.mask_2d_to_4d(mask, dtype, tgt_len=None)

prompt_strategies.pygmalion

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.pygmalion.html

Contents:

  • prompt_strategies.pygmalion
  • Classes
    • PygmalionPromptTokenizingStrategy
    • PygmalionPrompter

prompt_strategies.pygmalion

Module containing the PygmalionPromptTokenizingStrategy and PygmalionPrompter class

Tokenizing strategy for Pygmalion.

Prompter for Pygmalion.

Examples:

Example 1 (python):

prompt_strategies.pygmalion.PygmalionPromptTokenizingStrategy(
    prompter,
    tokenizer,
    *args,
    **kwargs,
)

Example 2 (python):

prompt_strategies.pygmalion.PygmalionPrompter(*args, **kwargs)

utils.callbacks.mlflow_

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.mlflow_.html

Contents:

  • utils.callbacks.mlflow_
  • Classes
    • SaveAxolotlConfigtoMlflowCallback

utils.callbacks.mlflow_

MLFlow module for trainer callbacks

Callback to save axolotl config to mlflow

Examples:

Example 1 (python):

utils.callbacks.mlflow_.SaveAxolotlConfigtoMlflowCallback(axolotl_config_path)

loaders.adapter

URL: https://docs.axolotl.ai/docs/api/loaders.adapter.html

Contents:

  • loaders.adapter
  • Functions
    • setup_quantized_meta_for_peft
    • setup_quantized_peft_meta_for_training

Adapter loading functionality, including LoRA / QLoRA and associated utils

Replaces quant_state.to with a dummy function to prevent PEFT from moving quant_state to meta device

Replaces dummy quant_state.to method with the original function to allow training to continue

Examples:

Example 1 (python):

loaders.adapter.setup_quantized_meta_for_peft(model)

Example 2 (python):

loaders.adapter.setup_quantized_peft_meta_for_training(model)

cli.cloud.base

URL: https://docs.axolotl.ai/docs/api/cli.cloud.base.html

Contents:

  • cli.cloud.base
  • Classes
    • Cloud

base class for cloud platforms from cli

Abstract base class for cloud platforms.

Examples:

Example 1 (python):

cli.cloud.base.Cloud()

monkeypatch.llama_attn_hijack_flash

URL: https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_flash.html

Contents:

  • monkeypatch.llama_attn_hijack_flash
  • Functions
    • flashattn_forward_with_s2attn

monkeypatch.llama_attn_hijack_flash

Flash attention monkey patch for llama model

Input shape: Batch x Time x Channel

From: https://github.com/dvlab-research/LongLoRA/blob/main/llama_attn_replace.py

attention_mask: [bsz, q_len]

cu_seqlens will be ignored if provided max_seqlen will be ignored if provided

Examples:

Example 1 (python):

monkeypatch.llama_attn_hijack_flash.flashattn_forward_with_s2attn(
    self,
    hidden_states,
    attention_mask=None,
    position_ids=None,
    past_key_value=None,
    output_attentions=False,
    use_cache=False,
    padding_mask=None,
    cu_seqlens=None,
    max_seqlen=None,
)

monkeypatch.llama_patch_multipack

URL: https://docs.axolotl.ai/docs/api/monkeypatch.llama_patch_multipack.html

Contents:

  • monkeypatch.llama_patch_multipack

monkeypatch.llama_patch_multipack

Patched LlamaAttention to use torch.nn.functional.scaled_dot_product_attention


cli.inference

URL: https://docs.axolotl.ai/docs/api/cli.inference.html

Contents:

  • cli.inference
  • Functions
    • do_cli
      • Parameters
    • do_inference
      • Parameters
    • do_inference_gradio
      • Parameters
    • get_multi_line_input
      • Returns

CLI to run inference on a trained model.

Parses axolotl config, CLI args, and calls do_inference or do_inference_gradio.

Runs inference on the command line in a loop. User input is accepted, a chat template is (optionally) applied, and the model specified in the axolotl config is used to generate completions according to a default generation config.

Runs inference in a Gradio interface. User input is accepted, a chat template is (optionally) applied, and the model specified in the axolotl config is used to generate completions according to a default generation config.

Gets multi-line input from terminal.

Examples:

Example 1 (python):

cli.inference.do_cli(config=Path('examples/'), gradio=False, **kwargs)

Example 2 (python):

cli.inference.do_inference(cfg, cli_args)

Example 3 (python):

cli.inference.do_inference_gradio(cfg, cli_args)

Example 4 (python):

cli.inference.get_multi_line_input()

loaders.tokenizer

URL: https://docs.axolotl.ai/docs/api/loaders.tokenizer.html

Contents:

  • loaders.tokenizer
  • Functions
    • load_tokenizer
    • modify_tokenizer_files
      • Parameters
      • Returns

Tokenizer loading functionality and associated utils

Load and configure the tokenizer based on the provided config.

Modify tokenizer files to replace added_tokens strings, save to output directory, and return the path to the modified tokenizer.

This only works with reserved tokens that were added to the tokenizer, not tokens already part of the vocab.

Ref: https://github.com/huggingface/transformers/issues/27974#issuecomment-1854188941

Examples:

Example 1 (python):

loaders.tokenizer.load_tokenizer(cfg)

Example 2 (python):

loaders.tokenizer.modify_tokenizer_files(
    tokenizer_path,
    token_mappings,
    output_dir,
)

cli.utils.sweeps

URL: https://docs.axolotl.ai/docs/api/cli.utils.sweeps.html

Contents:

  • cli.utils.sweeps
  • Functions
    • generate_sweep_configs
      • Parameters
      • Returns
      • Example

Utilities for handling sweeps over configs for axolotl train CLI command

Recursively generates all possible configurations by applying sweeps to the base config.

sweeps_config = { learning_rate: [0.1, 0.01], _: [ {load_in_8bit: True, adapter: lora}, {load_in_4bit: True, adapter: qlora} ] }

Examples:

Example 1 (python):

cli.utils.sweeps.generate_sweep_configs(base_config, sweeps_config)

prompt_strategies.dpo.chatml

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chatml.html

Contents:

  • prompt_strategies.dpo.chatml
  • Functions
    • argilla_chat
    • icr
    • intel
    • ultra

prompt_strategies.dpo.chatml

DPO strategies for chatml

for argilla/dpo-mix-7k conversations

chatml transforms for datasets with system, input, chosen, rejected ex. https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs

For Intel Orca DPO Pairs

for ultrafeedback binarized conversations

Examples:

Example 1 (python):

prompt_strategies.dpo.chatml.argilla_chat(cfg, **kwargs)

Example 2 (python):

prompt_strategies.dpo.chatml.icr(cfg, **kwargs)

Example 3 (python):

prompt_strategies.dpo.chatml.intel(cfg, **kwargs)

Example 4 (python):

prompt_strategies.dpo.chatml.ultra(cfg, **kwargs)

cli.quantize

URL: https://docs.axolotl.ai/docs/api/cli.quantize.html

Contents:

  • cli.quantize
  • Functions
    • do_quantize
      • Parameters

CLI to post-training quantize a model using torchao

Quantizes a models models weights

Examples:

Example 1 (python):

cli.quantize.do_quantize(config, cli_args)

utils.dict

URL: https://docs.axolotl.ai/docs/api/utils.dict.html

Contents:

  • utils.dict
  • Classes
    • DictDefault
  • Functions
    • remove_none_values

Module containing the DictDefault class

A Dict that returns None instead of returning empty Dict for missing keys.

Remove null from a dictionary-like obj or list. These can appear due to Dataset loading causing schema merge. See https://github.com/axolotl-ai-cloud/axolotl/pull/2909

Examples:

Example 1 (python):

utils.dict.DictDefault()

Example 2 (python):

utils.dict.remove_none_values(obj)

API Reference

URL: https://docs.axolotl.ai/docs/api/

Contents:

  • API Reference
  • Core
  • CLI
  • Trainers
  • Model Loading
  • Mixins
  • Context Managers
  • Prompt Strategies
  • Kernels
  • Monkey Patches

Core functionality for training

Command-line interface

Training implementations

Functionality for loading and patching models, tokenizers, etc.

Mixin classes for augmenting trainers

Context managers for altering trainer behaviors

Prompt formatting strategies

Low-level performance optimizations

Runtime patches for model optimizations

Pydantic data models for Axolotl config

Third-party integrations and extensions

Common utilities and shared functionality

Custom model implementations

Data processing utilities


monkeypatch.lora_kernels

URL: https://docs.axolotl.ai/docs/api/monkeypatch.lora_kernels.html

Contents:

  • monkeypatch.lora_kernels
  • Classes
    • FakeMLP
  • Functions
    • apply_lora_kernel_patches
      • Parameters
      • Returns
      • Raises
      • Note
    • get_attention_cls_from_config

monkeypatch.lora_kernels

Module for patching custom LoRA Triton kernels and torch.autograd functions.

placeholder MLP for triton patching

Applies optimized Triton kernel patches to a PEFT model.

Patches a PEFT model with optimized implementations for MLP and attention computations. The optimizations include custom Triton kernels for activation functions and specialized autograd functions for LoRA computations.

The optimizations require LoRA adapters with no dropout and no bias terms. The function will skip patching if these conditions arent met.

Get the appropriate attention class by inspecting the model config. Uses dynamic import to support any model architecture that follows the standard transformers naming convention.

Get the layers of the model. Handles text-only and multimodal models.

Original implementation of output projection without optimizations.

Original implementation of QKV projection without optimizations.

Given an axolotl config, this method patches the inferred attention class forward pass with optimized LoRA implementations.

It modifies the attention class to use optimized QKV and output projections. The original implementation is preserved and can be restored if needed.

Examples:

Example 1 (python):

monkeypatch.lora_kernels.FakeMLP(gate_proj, up_proj, down_proj)

Example 2 (python):

monkeypatch.lora_kernels.apply_lora_kernel_patches(model, cfg)

Example 3 (python):

monkeypatch.lora_kernels.get_attention_cls_from_config(cfg)

Example 4 (python):

monkeypatch.lora_kernels.get_layers(model)

monkeypatch.stablelm_attn_hijack_flash

URL: https://docs.axolotl.ai/docs/api/monkeypatch.stablelm_attn_hijack_flash.html

Contents:

  • monkeypatch.stablelm_attn_hijack_flash
  • Functions
    • repeat_kv
    • rotate_half

monkeypatch.stablelm_attn_hijack_flash

PyTorch StableLM Epoch model.

This is the equivalent of torch.repeat_interleave(x, dim=1, repeats=n_rep). The hidden states go from (batch, num_key_value_heads, seqlen, head_dim) to (batch, num_attention_heads, seqlen, head_dim)

Rotates half the hidden dims of the input.

Examples:

Example 1 (python):

monkeypatch.stablelm_attn_hijack_flash.repeat_kv(hidden_states, n_rep)

Example 2 (python):

monkeypatch.stablelm_attn_hijack_flash.rotate_half(x)

core.trainers.mixins.rng_state_loader

URL: https://docs.axolotl.ai/docs/api/core.trainers.mixins.rng_state_loader.html

Contents:

  • core.trainers.mixins.rng_state_loader
  • Classes
    • RngLoaderMixin

core.trainers.mixins.rng_state_loader

Temporary fix/override for bug in resume from checkpoint

See https://github.com/huggingface/transformers/pull/37162

TODO: Remove when upstream added PR to release

mixin for method override to load RNG states from a checkpoint

Examples:

Example 1 (python):

core.trainers.mixins.rng_state_loader.RngLoaderMixin()

core.trainers.utils

URL: https://docs.axolotl.ai/docs/api/core.trainers.utils.html

Contents:

  • core.trainers.utils

Utils for Axolotl trainers


core.training_args

URL: https://docs.axolotl.ai/docs/api/core.training_args.html

Contents:

  • core.training_args
  • Classes
    • AxolotlCPOConfig
    • AxolotlKTOConfig
    • AxolotlORPOConfig
    • AxolotlPRMConfig
    • AxolotlRewardConfig
    • AxolotlTrainingArguments

extra axolotl specific training args

CPO config for CPO training

KTO config for KTO training

ORPO config for ORPO training

PRM config for PRM training

Reward config for Reward training

Training arguments for Causal trainer

This code is duplicated due to HF TrainingArguments not setting output_dir with a default value so it cant be used as a mixin.

Examples:

Example 1 (python):

core.training_args.AxolotlCPOConfig(simpo_gamma=None)

Example 2 (python):

core.training_args.AxolotlKTOConfig()

Example 3 (python):

core.training_args.AxolotlORPOConfig()

Example 4 (python):

core.training_args.AxolotlPRMConfig()

monkeypatch.btlm_attn_hijack_flash

URL: https://docs.axolotl.ai/docs/api/monkeypatch.btlm_attn_hijack_flash.html

Contents:

  • monkeypatch.btlm_attn_hijack_flash

monkeypatch.btlm_attn_hijack_flash

Flash attention monkey patch for cerebras btlm model


prompt_strategies.dpo.passthrough

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.passthrough.html

Contents:

  • prompt_strategies.dpo.passthrough

prompt_strategies.dpo.passthrough

DPO prompt strategies passthrough/zero-processing strategy


kernels.swiglu

URL: https://docs.axolotl.ai/docs/api/kernels.swiglu.html

Contents:

  • kernels.swiglu
  • Functions
    • swiglu_backward
      • Parameters
      • Returns
    • swiglu_forward
      • Parameters
      • Returns

Module for definition of SwiGLU Triton kernels.

See “GLU Variants Improve Transformer” (https://arxiv.org/abs/2002.05202).

Credit to unsloth (https://unsloth.ai/) for inspiration for this implementation.

SwiGLU backward pass using in-place operations.

SwiGLU forward pass. Computes SwiGLU activation: x * sigmoid(x) * up, where x is the gate tensor.

Examples:

Example 1 (python):

kernels.swiglu.swiglu_backward(grad_output, gate, up)

Example 2 (python):

kernels.swiglu.swiglu_forward(gate, up)

core.trainers.grpo.trainer

URL: https://docs.axolotl.ai/docs/api/core.trainers.grpo.trainer.html

Contents:

  • core.trainers.grpo.trainer
  • Classes
    • AxolotlGRPOSequenceParallelTrainer
      • Methods
        • get_train_dataloader
    • AxolotlGRPOTrainer

core.trainers.grpo.trainer

Axolotl GRPO trainers (with and without sequence parallelism handling)

Extend the base GRPOTrainer for sequence parallelism handling

Get dataloader for training

Extend the base GRPOTrainer for axolotl helpers

Examples:

Example 1 (python):

core.trainers.grpo.trainer.AxolotlGRPOSequenceParallelTrainer(
    model,
    reward_funcs,
    args=None,
    train_dataset=None,
    eval_dataset=None,
    processing_class=None,
    reward_processing_classes=None,
    callbacks=None,
    optimizers=(None, None),
    peft_config=None,
    optimizer_cls_and_kwargs=None,
)

Example 2 (python):

core.trainers.grpo.trainer.AxolotlGRPOSequenceParallelTrainer.get_train_dataloader(
)

Example 3 (python):

core.trainers.grpo.trainer.AxolotlGRPOTrainer(*args, **kwargs)

prompt_strategies.user_defined

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.user_defined.html

Contents:

  • prompt_strategies.user_defined
  • Classes
    • UserDefinedDatasetConfig
    • UserDefinedPromptTokenizationStrategy

prompt_strategies.user_defined

User Defined prompts with configuration from the YML config

dataclass configuration representing a userdefined dataset type

Prompt Tokenization Strategy for user defined prompts

Examples:

Example 1 (python):

prompt_strategies.user_defined.UserDefinedDatasetConfig(
    system_prompt='',
    field_system='system',
    field_instruction='instruction',
    field_input='input',
    field_output='output',
    format='{instruction} {input} ',
    no_input_format='{instruction} ',
    system_format='{system}',
)

Example 2 (python):

prompt_strategies.user_defined.UserDefinedPromptTokenizationStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

utils.schemas.training

URL: https://docs.axolotl.ai/docs/api/utils.schemas.training.html

Contents:

  • utils.schemas.training
  • Classes
    • HyperparametersConfig
    • JaggedLRConfig
    • LrGroup

utils.schemas.training

Pydantic models for training hyperparameters

Training hyperparams configuration subset

JaggedLR configuration subset, can be used w/ ReLoRA training

Custom learning rate group configuration

Examples:

Example 1 (python):

utils.schemas.training.HyperparametersConfig()

Example 2 (python):

utils.schemas.training.JaggedLRConfig()

Example 3 (python):

utils.schemas.training.LrGroup()

utils.quantization

URL: https://docs.axolotl.ai/docs/api/utils.quantization.html

Contents:

  • utils.quantization
  • Functions
    • convert_qat_model
    • get_quantization_config
      • Parameters
      • Returns
      • Raises
    • prepare_model_for_qat
      • Parameters
      • Raises

Utilities for quantization including QAT and PTQ using torchao.

This function converts a QAT model which has fake quantized layers back to the original model.

This function is used to build a post-training quantization config.

This function is used to prepare a model for QAT by swapping the models linear layers with fake quantized linear layers, and optionally the embedding weights with fake quantized embedding weights.

This function is used to quantize a model.

Examples:

Example 1 (python):

utils.quantization.convert_qat_model(model, quantize_embedding=False)

Example 2 (python):

utils.quantization.get_quantization_config(
    weight_dtype,
    activation_dtype=None,
    group_size=None,
)

Example 3 (python):

utils.quantization.prepare_model_for_qat(
    model,
    weight_dtype,
    group_size=None,
    activation_dtype=None,
    quantize_embedding=False,
)

Example 4 (python):

utils.quantization.quantize_model(
    model,
    weight_dtype,
    group_size=None,
    activation_dtype=None,
    quantize_embedding=None,
)

logging_config

URL: https://docs.axolotl.ai/docs/api/logging_config.html

Contents:

  • logging_config
  • Classes
    • AxolotlLogger
    • AxolotlOrWarnErrorFilter
    • ColorfulFormatter
  • Functions
    • configure_logging

Common logging module for axolotl.

Logger that applies filtering to non-axolotl loggers.

Allows ANY WARNING or higher (unless overridden by LOG_LEVEL). Allows axolotl.* at INFO or higher (unless overridden by AXOLOTL_LOG_LEVEL). Drops all other records (i.e. non-axolotl.INFO, DEBUG, etc. by default).

Formatter to add coloring to log messages by log type

Configure with default logging

Examples:

Example 1 (python):

logging_config.AxolotlLogger(name, level=logging.NOTSET)

Example 2 (python):

logging_config.AxolotlOrWarnErrorFilter(**kwargs)

Example 3 (python):

logging_config.ColorfulFormatter()

Example 4 (python):

logging_config.configure_logging()

prompt_strategies.stepwise_supervised

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.stepwise_supervised.html

Contents:

  • prompt_strategies.stepwise_supervised
  • Classes
    • StepwiseSupervisedPromptTokenizingStrategy

prompt_strategies.stepwise_supervised

Module for stepwise datasets, typically including a prompt and reasoning traces, and (optionally) per-step, or per-prompt-trace labels for reward modelling.

Tokenizing strategy for supervised stepwise datasets, typically used for COT-reasoning. These datasets should include the following columns: - prompt: the prompt text - completions: a list of n completion steps - labels: a list of n labels indicating the “correctness” of each step

Examples:

Example 1 (python):

prompt_strategies.stepwise_supervised.StepwiseSupervisedPromptTokenizingStrategy(
    tokenizer,
    sequence_len=2048,
    step_separator='\n',
    max_completion_length=None,
    train_on_last_step_only=False,
)

utils.schemas.model

URL: https://docs.axolotl.ai/docs/api/utils.schemas.model.html

Contents:

  • utils.schemas.model
  • Classes
    • ModelInputConfig
    • ModelOutputConfig
    • SpecialTokensConfig

Pydantic models for model input / output, etc. configuration

Model configuration subset

model save configuration subset

Special tokens configuration subset

Examples:

Example 1 (python):

utils.schemas.model.ModelInputConfig()

Example 2 (python):

utils.schemas.model.ModelOutputConfig()

Example 3 (python):

utils.schemas.model.SpecialTokensConfig()

utils.schemas.enums

URL: https://docs.axolotl.ai/docs/api/utils.schemas.enums.html

Contents:

  • utils.schemas.enums
  • Classes
    • ChatTemplate
    • CustomSupportedOptimizers
    • RLType
    • RingAttnFunc

Enums for Axolotl input config

Chat templates configuration subset

Custom supported optimizers

RL trainer type configuration subset

Enum class for supported ring-flash-attn implementations

Examples:

Example 1 (python):

utils.schemas.enums.ChatTemplate()

Example 2 (python):

utils.schemas.enums.CustomSupportedOptimizers()

Example 3 (python):

utils.schemas.enums.RLType()

Example 4 (python):

utils.schemas.enums.RingAttnFunc()

core.trainers.trl

URL: https://docs.axolotl.ai/docs/api/core.trainers.trl.html

Contents:

  • core.trainers.trl
  • Classes
    • AxolotlCPOTrainer
    • AxolotlKTOTrainer
    • AxolotlORPOTrainer
    • AxolotlPRMTrainer
    • AxolotlRewardTrainer

Module for TRL RL trainers

Extend the base CPOTrainer for axolotl helpers

Extend the base KTOTrainer for axolotl helpers

Extend the base ORPOTrainer for axolotl helpers

Extend the base trl.PRMTrainer for axolotl helpers

Extend the base RewardTrainer for axolotl helpers

Examples:

Example 1 (python):

core.trainers.trl.AxolotlCPOTrainer(*args, **kwargs)

Example 2 (python):

core.trainers.trl.AxolotlKTOTrainer(*args, **kwargs)

Example 3 (python):

core.trainers.trl.AxolotlORPOTrainer(*args, **kwargs)

Example 4 (python):

core.trainers.trl.AxolotlPRMTrainer(*args, **kwargs)

utils.schedulers

URL: https://docs.axolotl.ai/docs/api/utils.schedulers.html

Contents:

  • utils.schedulers
  • Classes
    • InterpolatingLogScheduler
    • JaggedLRRestartScheduler
    • RexLR
      • Parameters
  • Functions
    • get_cosine_schedule_with_min_lr
      • Create a learning rate schedule which has
    • get_cosine_schedule_with_quadratic_warmup

Module for custom LRScheduler class

A scheduler that interpolates learning rates in a logarithmic fashion

Wraps another scheduler to apply per-lora-restart learning rate warmups.

Reflected Exponential (REX) learning rate scheduler.

Create a schedule with a learning rate that decreases following the values of the cosine function between the initial lr set in the optimizer to 0, after a warmup period during which it increases linearly between 0 and the initial lr set in the optimizer.

torch.optim.lr_scheduler.LambdaLR with the appropriate schedule.

Implementation of Continual Pre-Training of Large Language Models: How to (re)warm your model? (https://arxiv.org/pdf/2308.04014.pdf) Create a schedule with a learning rate that decreases following the values of the cosine function between the initial lr set in the optimizer to min_lr_ratio until num_training_steps * constant_lr_ratio, after constant_rate returns constant value of min_rate , after a warmup period during which it increases linearly between 0 and the initial lr set in the optimizer.

torch.optim.lr_scheduler.LambdaLR with the appropriate schedule.

Examples:

Example 1 (python):

utils.schedulers.InterpolatingLogScheduler(
    optimizer,
    num_steps,
    min_lr,
    max_lr,
    last_epoch=-1,
)

Example 2 (python):

utils.schedulers.JaggedLRRestartScheduler(
    optimizer,
    inner_schedule,
    jagged_restart_steps,
    jagged_restart_warmup_steps,
    jagged_restart_anneal_steps=1,
    min_lr_scale=0.001,
)

Example 3 (python):

utils.schedulers.RexLR(
    optimizer,
    max_lr,
    min_lr,
    total_steps=0,
    num_warmup_steps=0,
    last_step=0,
)

Example 4 (python):

utils.schedulers.get_cosine_schedule_with_min_lr(
    optimizer,
    num_warmup_steps,
    num_training_steps,
    min_lr_ratio=0.0,
)

cli.merge_lora

URL: https://docs.axolotl.ai/docs/api/cli.merge_lora.html

Contents:

  • cli.merge_lora
  • Functions
    • do_cli
      • Parameters
      • Raises
    • do_merge_lora
      • Parameters

CLI to merge a trained LoRA into a base model.

Parses axolotl config, CLI args, and calls do_merge_lora. Note that various config values will be overwritten to allow the LoRA merge logic to work as expected (load_in_8bit=False, load_in4bit=False, flash_attention=False, etc.).

Calls transformers merge_and_unload on the model given in the axolotl config along with the LoRA adapters to combine them into a single base model.

Examples:

Example 1 (python):

cli.merge_lora.do_cli(config=Path('examples/'), **kwargs)

Example 2 (python):

cli.merge_lora.do_merge_lora(cfg)

prompt_strategies.alpaca_w_system

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_w_system.html

Contents:

  • prompt_strategies.alpaca_w_system
  • Classes
    • InstructionWSystemPromptTokenizingStrategy
    • OpenOrcaPromptTokenizingStrategy
    • OpenOrcaSystemDataPrompter
    • SystemDataPrompter

prompt_strategies.alpaca_w_system

Prompt strategies loader for alpaca instruction datasets with system prompts

Tokenizing strategy for instruction-based prompts.

Tokenizing strategy for OpenOrca datasets

Alpaca Style Prompter that uses system prompts from the dataset, with OpenOrca prompts

Alpaca Style Prompter that uses system prompts from the dataset

Examples:

Example 1 (python):

prompt_strategies.alpaca_w_system.InstructionWSystemPromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 2 (python):

prompt_strategies.alpaca_w_system.OpenOrcaPromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 3 (python):

prompt_strategies.alpaca_w_system.OpenOrcaSystemDataPrompter(
    prompt_style=PromptStyle.INSTRUCT.value,
)

Example 4 (python):

prompt_strategies.alpaca_w_system.SystemDataPrompter(
    prompt_style=PromptStyle.INSTRUCT.value,
)

loaders.patch_manager

URL: https://docs.axolotl.ai/docs/api/loaders.patch_manager.html

Contents:

  • loaders.patch_manager
  • Classes
    • PatchManager
      • Attributes
      • Methods
        • apply_post_model_load_patches
        • apply_post_plugin_pre_model_load_patches
        • apply_pre_model_load_patches

loaders.patch_manager

Patch manager class implementation to complement axolotl.loaders.ModelLoader.

Applies pre- and post-model load patches for various fixes and optimizations.

Manages the application of patches during the model loading process.

Apply patches that require the model instance.

Apply post plugin-pre_model_load load patches based on config.

Apply pre-model load patches based on config.

Examples:

Example 1 (python):

loaders.patch_manager.PatchManager(cfg, model_config, inference=False)

Example 2 (python):

loaders.patch_manager.PatchManager.apply_post_model_load_patches(model)

Example 3 (python):

loaders.patch_manager.PatchManager.apply_post_plugin_pre_model_load_patches()

Example 4 (python):

loaders.patch_manager.PatchManager.apply_pre_model_load_patches()

utils.schemas.peft

URL: https://docs.axolotl.ai/docs/api/utils.schemas.peft.html

Contents:

  • utils.schemas.peft
  • Classes
    • LoftQConfig
    • LoraConfig
    • PeftConfig
    • ReLoRAConfig

Pydantic models for PEFT-related configuration

LoftQ configuration subset

Peft / LoRA configuration subset

peftq configuration subset

ReLoRA configuration subset

Examples:

Example 1 (python):

utils.schemas.peft.LoftQConfig()

Example 2 (python):

utils.schemas.peft.LoraConfig()

Example 3 (python):

utils.schemas.peft.PeftConfig()

Example 4 (python):

utils.schemas.peft.ReLoRAConfig()

common.const

URL: https://docs.axolotl.ai/docs/api/common.const.html

Contents:

  • common.const

Various shared constants


prompt_strategies.kto.user_defined

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.kto.user_defined.html

Contents:

  • prompt_strategies.kto.user_defined

prompt_strategies.kto.user_defined

User-defined KTO strategies


prompt_strategies.base

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.base.html

Contents:

  • prompt_strategies.base

prompt_strategies.base

module for base dataset transform strategies


cli.delinearize_llama4

URL: https://docs.axolotl.ai/docs/api/cli.delinearize_llama4.html

Contents:

  • cli.delinearize_llama4
  • Functions
    • do_cli
      • Parameters

cli.delinearize_llama4

CLI tool to delinearize quantized/Linearized Llama-4 models.

Convert a patched HF format Llama4 model (with separated projections) back to the original HF format (with fused projections).

Examples:

Example 1 (python):

cli.delinearize_llama4.do_cli(model, output)

integrations.base

URL: https://docs.axolotl.ai/docs/api/integrations.base.html

Contents:

  • integrations.base
  • Classes
    • BaseOptimizerFactory
      • Methods
        • get_decay_parameter_names
    • BasePlugin
      • Note
      • Methods
        • add_callbacks_post_trainer
          • Parameters

Base class for all plugins.

A plugin is a reusable, modular, and self-contained piece of code that extends the functionality of Axolotl. Plugins can be used to integrate third-party models, modify the training process, or add new features.

To create a new plugin, you need to inherit from the BasePlugin class and implement the required methods.

Base class for factories to create custom optimizers

Get all parameter names that weight decay will be applied to.

This function filters out parameters in two ways: 1. By layer type (instances of layers specified in ALL_LAYERNORM_LAYERS) 2. By parameter name patterns (containing bias, or variation of norm)

Base class for all plugins. Defines the interface for plugin methods.

A plugin is a reusable, modular, and self-contained piece of code that extends the functionality of Axolotl. Plugins can be used to integrate third-party models, modify the training process, or add new features.

To create a new plugin, you need to inherit from the BasePlugin class and implement the required methods.

Plugin methods include: - register(cfg): Registers the plugin with the given configuration. - load_datasets(cfg): Loads and preprocesses the dataset for training. - pre_model_load(cfg): Performs actions before the model is loaded. - post_model_build(cfg, model): Performs actions after the model is loaded, but before LoRA adapters are applied. - pre_lora_load(cfg, model): Performs actions before LoRA weights are loaded. - post_lora_load(cfg, model): Performs actions after LoRA weights are loaded. - post_model_load(cfg, model): Performs actions after the model is loaded, inclusive of any adapters. - post_trainer_create(cfg, trainer): Performs actions after the trainer is created. - create_optimizer(cfg, trainer): Creates and returns an optimizer for training. - create_lr_scheduler(cfg, trainer, optimizer, num_training_steps): Creates and returns a learning rate scheduler. - add_callbacks_pre_trainer(cfg, model): Adds callbacks to the trainer before training. - add_callbacks_post_trainer(cfg, trainer): Adds callbacks to the trainer after training.

Adds callbacks to the trainer after creating the trainer. This is useful for callbacks that require access to the model or trainer.

Set up callbacks before creating the trainer.

Creates and returns a learning rate scheduler.

Creates and returns an optimizer for training.

Returns a custom class for the collator.

Returns a pydantic model for the plugins input arguments.

Returns a custom class for the trainer.

Returns custom training arguments to set on TrainingArgs.

Returns a dataclass model for the plugins training arguments.

Loads and preprocesses the dataset for training.

Performs actions after LoRA weights are loaded.

Performs actions after the model is built/loaded, but before any adapters are applied.

Performs actions after the model is loaded.

Performs actions after training is complete.

Performs actions after training is complete and the model is unloaded.

Performs actions after the trainer is created.

Performs actions before LoRA weights are loaded.

Performs actions before the model is loaded.

Registers the plugin with the given configuration as an unparsed dict.

The PluginManager class is responsible for loading and managing plugins. It should be a singleton so it can be accessed from anywhere in the codebase.

Key methods include: - get_instance(): Static method to get the singleton instance of PluginManager. - register(plugin_name: str): Registers a new plugin by its name. - pre_model_load(cfg): Calls the pre_model_load method of all registered plugins.

Calls the add_callbacks_post_trainer method of all registered plugins.

Calls the add_callbacks_pre_trainer method of all registered plugins.

Calls the create_lr_scheduler method of all registered plugins and returns the first non-None scheduler.

Calls the create_optimizer method of all registered plugins and returns the first non-None optimizer.

Calls the get_collator_cls_and_kwargs method of all registered plugins and returns the first non-None collator class.

Parameters: cfg (dict): The configuration for the plugins. is_eval (bool): Whether this is an eval split.

Returns: object: The collator class, or None if none was found.

Returns a list of Pydantic classes for all registered plugins input arguments.

Returns the singleton instance of PluginManager. If the instance doesnt exist, it creates a new one.

Calls the get_trainer_cls method of all registered plugins and returns the first non-None trainer class.

Calls the get_training_args method of all registered plugins and returns the combined training arguments.

Parameters: cfg (dict): The configuration for the plugins.

Returns: object: The training arguments

Returns a list of dataclasses for all registered plugins training args mixins

Returns: list[str]: A list of dataclsses

Calls the load_datasets method of each registered plugin.

Calls the post_lora_load method of all registered plugins.

Calls the post_model_build method of all registered plugins after the model has been built / loaded, but before any adapters have been applied.

Calls the post_model_load method of all registered plugins after the model has been loaded inclusive of any adapters.

Calls the post_train method of all registered plugins.

Calls the post_train_unload method of all registered plugins.

Calls the post_trainer_create method of all registered plugins.

Calls the pre_lora_load method of all registered plugins.

Calls the pre_model_load method of all registered plugins.

Registers a new plugin by its name.

Loads a plugin based on the given plugin name.

The plugin name should be in the format “module_name.class_name”. This function splits the plugin name into module and class, imports the module, retrieves the class from the module, and creates an instance of the class.

Examples:

Example 1 (python):

integrations.base.BaseOptimizerFactory()

Example 2 (python):

integrations.base.BaseOptimizerFactory.get_decay_parameter_names(model)

Example 3 (python):

integrations.base.BasePlugin()

Example 4 (python):

integrations.base.BasePlugin.add_callbacks_post_trainer(cfg, trainer)

prompt_strategies.chat_template

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.chat_template.html

Contents:

  • prompt_strategies.chat_template
  • Classes
    • ChatTemplatePrompter
      • Methods
        • build_prompt
          • Parameters
    • ChatTemplateStrategy
      • Methods
        • find_first_eot_token
        • find_turn

prompt_strategies.chat_template

HF Chat Templates prompt strategy

Prompter for HF chat templates

Build a prompt from a conversation.

Tokenizing strategy for instruction-based prompts.

Find the first EOT token in the input_ids starting from start_idx.

Locate the starting and ending indices of the specified turn in a conversation.

Public method that can handle either a single prompt or a batch of prompts.

Mistral prompter for chat template.

Mistral strategy for chat template.

Find the first EOT token in the input_ids starting from start_idx.

Load chat template strategy based on configuration.

Examples:

Example 1 (python):

prompt_strategies.chat_template.ChatTemplatePrompter(
    tokenizer,
    chat_template,
    processor=None,
    max_length=2048,
    message_property_mappings=None,
    message_field_training=None,
    message_field_training_detail=None,
    field_messages='messages',
    field_system='system',
    field_tools='tools',
    field_thinking='reasoning_content',
    roles=None,
    template_thinking_key='reasoning_content',
    chat_template_kwargs=None,
    drop_system_message=False,
)

Example 2 (python):

prompt_strategies.chat_template.ChatTemplatePrompter.build_prompt(
    conversation,
    add_generation_prompt=False,
    images=None,
    tools=None,
)

Example 3 (python):

prompt_strategies.chat_template.ChatTemplateStrategy(
    prompter,
    tokenizer,
    train_on_inputs,
    sequence_len,
    roles_to_train=None,
    train_on_eos=None,
    train_on_eot=None,
    eot_tokens=None,
    split_thinking=False,
)

Example 4 (python):

prompt_strategies.chat_template.ChatTemplateStrategy.find_first_eot_token(
    input_ids,
    start_idx,
)

kernels.quantize

URL: https://docs.axolotl.ai/docs/api/kernels.quantize.html

Contents:

  • kernels.quantize
  • Functions
    • dequantize
      • Parameters
      • Returns
      • Raises
      • Note

Dequantization utilities for bitsandbytes integration.

Fast NF4 dequantization using bitsandbytes CUDA kernels.

Performs efficient dequantization of weights from NF4 format using bitsandbytes optimized CUDA implementations. Supports both legacy list and new QuantState formats.

Uses CUDA streams for better performance when available in newer bitsandbytes versions (>0.43.3).

Examples:

Example 1 (python):

kernels.quantize.dequantize(W, quant_state=None, out=None)

integrations.spectrum.args

URL: https://docs.axolotl.ai/docs/api/integrations.spectrum.args.html

Contents:

  • integrations.spectrum.args
  • Classes
    • SpectrumArgs

integrations.spectrum.args

Module for handling Spectrum input arguments.

Input args for Spectrum.

Examples:

Example 1 (python):

integrations.spectrum.args.SpectrumArgs()

prompt_strategies.alpaca_chat

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_chat.html

Contents:

  • prompt_strategies.alpaca_chat
  • Classes
    • AlpacaChatPrompter
    • AlpacaConcisePrompter
    • AlpacaQAPromptTokenizingStrategy
    • CamelAIPromptTokenizingStrategy
    • NoSystemPrompter

prompt_strategies.alpaca_chat

Module for Alpaca prompt strategy classes

Alpaca Chat Prompter extending the system prompt to for chat-instruct answers

Alpaca Prompter extending the system prompt to ask for concise chat-instruct answers

Tokenizing strategy for AlpacaQA

Tokenizing strategy for CamelAI datasets

Null Prompter with no system prompts

Examples:

Example 1 (python):

prompt_strategies.alpaca_chat.AlpacaChatPrompter()

Example 2 (python):

prompt_strategies.alpaca_chat.AlpacaConcisePrompter(
    prompt_style=PromptStyle.INSTRUCT.value,
)

Example 3 (python):

prompt_strategies.alpaca_chat.AlpacaQAPromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 4 (python):

prompt_strategies.alpaca_chat.CamelAIPromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

utils.collators.mamba

URL: https://docs.axolotl.ai/docs/api/utils.collators.mamba.html

Contents:

  • utils.collators.mamba
  • Classes
    • MambaDataCollator

utils.collators.mamba

Collator for State Space Models (Mamba)

Examples:

Example 1 (python):

utils.collators.mamba.MambaDataCollator(tokenizer)

prompt_strategies.messages.chat

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.messages.chat.html

Contents:

  • prompt_strategies.messages.chat
  • Classes
    • ChatMessageDatasetWrappingStrategy

prompt_strategies.messages.chat

Chat dataset wrapping strategy for new internal messages representations

Chat dataset wrapping strategy for new internal messages representations

Examples:

Example 1 (python):

prompt_strategies.messages.chat.ChatMessageDatasetWrappingStrategy(
    processor,
    message_transform=None,
    formatter=None,
    **kwargs,
)

train

URL: https://docs.axolotl.ai/docs/api/train.html

Contents:

  • train
  • Functions
    • create_model_card
      • Parameters
    • execute_training
      • Parameters
    • handle_untrained_tokens_fix
      • Parameters
    • save_initial_configs
      • Parameters

Prepare and train a model on a dataset. Can also infer from a model or merge lora

Create a model card for the trained model if needed.

Execute the training process with appropriate SDP kernel configurations.

Apply fixes for untrained tokens if configured.

Save initial configurations before training.

Save the trained model according to configuration and training setup.

Load the tokenizer, processor (for multimodal models), and model based on configuration.

Load model, tokenizer, trainer, etc. Helper function to encapsulate the full trainer setup.

Set up the Axolotl badge and add the Axolotl config to the model card if available.

Set up the reference model for RL training if needed.

Set up signal handler for graceful termination.

Train a model on the given dataset.

Examples:

Example 1 (python):

train.create_model_card(cfg, trainer)

Example 2 (python):

train.execute_training(cfg, trainer, resume_from_checkpoint)

Example 3 (python):

train.handle_untrained_tokens_fix(
    cfg,
    model,
    tokenizer,
    train_dataset,
    safe_serialization,
)

Example 4 (python):

train.save_initial_configs(cfg, tokenizer, model, peft_config, processor)

cli.utils.load

URL: https://docs.axolotl.ai/docs/api/cli.utils.load.html

Contents:

  • cli.utils.load
  • Functions
    • load_model_and_tokenizer
      • Parameters
      • Returns

Utilities for model, tokenizer, etc. loading.

Helper function for loading a model, tokenizer, and processor specified in the given axolotl config.

Examples:

Example 1 (python):

cli.utils.load.load_model_and_tokenizer(cfg, inference=False)

loaders.model

URL: https://docs.axolotl.ai/docs/api/loaders.model.html

Contents:

  • loaders.model
  • Classes
    • ModelLoader
      • The loading process includes
      • Attributes
      • Methods
        • load
          • Returns

Model loader class implementation for loading, configuring, and patching various models.

Manages model configuration, initialization and application of patches during model loading.

This class orchestrates the entire process of loading a model from configuration to final preparation. It handles device mapping, quantization, attention mechanisms, adapter integration, and various optimizations.

Load and prepare the model with all configurations and patches.

Examples:

Example 1 (python):

loaders.model.ModelLoader(
    cfg,
    tokenizer,
    *,
    inference=False,
    reference_model=False,
    **kwargs,
)

Example 2 (python):

loaders.model.ModelLoader.load()

utils.distributed

URL: https://docs.axolotl.ai/docs/api/utils.distributed.html

Contents:

  • utils.distributed
  • Functions
    • barrier
    • cleanup_distributed
    • compute_and_broadcast
    • gather_from_all_ranks
    • gather_scalar_from_all_ranks
    • is_distributed
    • is_main_process
      • Returns

Utilities for distributed functionality.

Acts as a barrier to wait for all processes. This ensures that all processes reach the barrier before proceeding further.

Destroy process group if torch distributed is initialized. Called in training early termination or when training successfully completes.

Compute a value using the function fn only on the specified rank (default is 0). The value is then broadcasted to all other ranks.

Args: - fn (callable): A function that computes the value. This should not have any side effects. - rank (int, optional): The rank that computes the value. Default is 0.

Returns: - The computed value (int or float).

Run a callable fn on all ranks and gather the results on the specified rank.

Args: - fn (callable): A function that computes the value. This should not have any side effects. - rank (int, optional): The rank that gathers the values. Default is 0. - world_size (int, optional): Total number of processes in the current distributed setup.

Returns: - A list of computed values from all ranks if on the gathering rank, otherwise None.

Run a callable fn on all ranks and gather the results on the specified rank.

Args: - fn (callable): A function that computes the value. This should not have any side effects. - rank (int, optional): The rank that gathers the values. Default is 0. - world_size (int, optional): Total number of processes in the current distributed setup.

Returns: - A list of computed values from all ranks if on the gathering rank, otherwise None.

Check if distributed training is initialized.

Check if the current process is the main process. If not in distributed mode, always return True.

We use a simpler logic when the distributed state is not initialized: we just log on the 0-th local rank.

Run a callable fn1 on all ranks, gather the results, reduce them using fn2, and then broadcast the reduced result to all ranks.

Args: - fn1 (callable): A function that computes the value on each rank. - fn2 (callable): A reduction function that takes a list of values and returns a single value. - world_size (int, optional): Total number of processes in the current distributed setup.

Returns: - The reduced and broadcasted value.

runs the wrapped context so that rank 0 runs first before other ranks

Examples:

Example 1 (python):

utils.distributed.barrier()

Example 2 (python):

utils.distributed.cleanup_distributed()

Example 3 (python):

utils.distributed.compute_and_broadcast(fn)

Example 4 (python):

utils.distributed.gather_from_all_ranks(fn, world_size=1)

cli.config

URL: https://docs.axolotl.ai/docs/api/cli.config.html

Contents:

  • cli.config
  • Functions
    • check_remote_config
      • Parameters
      • Returns
      • Raises
    • choose_config
      • Parameters
      • Returns
      • Raises

Configuration loading and processing.

First, determines if the passed config is a valid HTTPS URL. Then, attempts to query for it and parse its content, first as JSON, then as YAML (YAML is preferred). Finally, the parsed content is written to a local file and its path is returned.

Helper method for choosing a axolotl config YAML file (considering only files ending with .yml or .yaml). If more than one config file exists in the passed path, the user is prompted to choose one.

Loads the axolotl configuration stored at config, validates it, and performs various setup.

Registers the plugins for the given configuration.

Examples:

Example 1 (python):

cli.config.check_remote_config(config)

Example 2 (python):

cli.config.choose_config(path)

Example 3 (python):

cli.config.load_cfg(config=Path('examples/'), **kwargs)

Example 4 (python):

cli.config.prepare_plugins(cfg)

cli.checks

URL: https://docs.axolotl.ai/docs/api/cli.checks.html

Contents:

  • cli.checks
  • Functions
    • check_accelerate_default_config
    • check_user_token
      • Returns
      • Raises

Various checks for Axolotl CLI.

Logs at warning level if no accelerate config file is found.

Checks for HF user info. Check is skipped if HF_HUB_OFFLINE=1.

Examples:

Example 1 (python):

cli.checks.check_accelerate_default_config()

Example 2 (python):

cli.checks.check_user_token()

prompt_strategies.llama2_chat

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.llama2_chat.html

Contents:

  • prompt_strategies.llama2_chat
  • Classes
    • LLama2ChatTokenizingStrategy
    • Llama2ChatConversation
      • Methods
        • append_message
        • get_prompt
    • Llama2ChatPrompter

prompt_strategies.llama2_chat

Prompt Strategy for finetuning Llama2 chat models see also 6c7fe27657/llama/generation.py (L213) for ma reference implementation.

This implementation is based on the Vicuna PR and the fastchat repo, see also: cdd7730686/fastchat/conversation.py (L847)

Use dataset type: “llama2_chat” in config.yml to use this prompt style.

E.g. in the config.yml:

The dataset itself should look like this:

in a jsonl file. The first message should be from the human, the second from gpt. For a custom system message, the first “from” can be “system” (followed by alternating “human” and “gpt” turns).

Important: Dont use “special_tokens:” in your config.yml if you are not sure what you are doing!

Tokenizing strategy for Llama2 prompts. adapted from https://github.com/lm-sys/FastChat/blob/main/fastchat/train/train.py

A class that manages prompt templates and keeps all conversation history. copied from https://github.com/lm-sys/FastChat/blob/main/fastchat/conversation.py

Append a new message.

Get the prompt for generation.

A prompter that generates prompts for Llama2 models.

Examples:

Example 1 (unknown):

datasets:
  - path: llama_finetune_train.jsonl
    type: llama2_chat

Example 2 (unknown):

{'conversations':[{"from": "human", "value": "Who are you?"}, {"from": "gpt", "value": "I am Vicuna"},...]}

Example 3 (python):

prompt_strategies.llama2_chat.LLama2ChatTokenizingStrategy(*args, **kwargs)

Example 4 (python):

prompt_strategies.llama2_chat.Llama2ChatConversation(
    name='llama2',
    system="[INST] <<SYS>>\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.\n<</SYS>>\n\n",
    roles=('[INST]', '[/INST]'),
    messages=list(),
    offset=0,
)

cli.utils

URL: https://docs.axolotl.ai/docs/api/cli.utils.html

Contents:

  • cli.utils

Init for axolotl.cli.utils module.


cli.utils.args

URL: https://docs.axolotl.ai/docs/api/cli.utils.args.html

Contents:

  • cli.utils.args
  • Functions
    • add_options_from_config
      • Parameters
      • Returns
    • add_options_from_dataclass
      • Parameters
      • Returns
    • filter_none_kwargs
      • Parameters

Utilities for axolotl CLI args.

Create Click options from the fields of a Pydantic model.

Create Click options from the fields of a dataclass.

Wraps function to remove None-valued kwargs.

Examples:

Example 1 (python):

cli.utils.args.add_options_from_config(config_class)

Example 2 (python):

cli.utils.args.add_options_from_dataclass(config_class)

Example 3 (python):

cli.utils.args.filter_none_kwargs(func)

integrations.grokfast.optimizer

URL: https://docs.axolotl.ai/docs/api/integrations.grokfast.optimizer.html

Contents:

  • integrations.grokfast.optimizer

integrations.grokfast.optimizer


core.builders.causal

URL: https://docs.axolotl.ai/docs/api/core.builders.causal.html

Contents:

  • core.builders.causal
  • Classes
    • HFCausalTrainerBuilder

Builder for causal trainers

Build the HuggingFace training args/trainer for causal models and reward modeling using TRL.

Examples:

Example 1 (python):

core.builders.causal.HFCausalTrainerBuilder(
    cfg,
    model,
    tokenizer,
    processor=None,
)

prompt_strategies.dpo.user_defined

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.user_defined.html

Contents:

  • prompt_strategies.dpo.user_defined

prompt_strategies.dpo.user_defined

User-defined DPO strategies


cli.evaluate

URL: https://docs.axolotl.ai/docs/api/cli.evaluate.html

Contents:

  • cli.evaluate
  • Functions
    • do_cli
      • Parameters
    • do_evaluate
      • Parameters

CLI to run evaluation on a model.

Parses axolotl config, CLI args, and calls do_evaluate.

Evaluates a transformers model by first loading the dataset(s) specified in the axolotl config, and then calling axolotl.evaluate.evaluate, which computes evaluation metrics on the given dataset(s) and writes them to disk.

Examples:

Example 1 (python):

cli.evaluate.do_cli(config=Path('examples/'), **kwargs)

Example 2 (python):

cli.evaluate.do_evaluate(cfg, cli_args)

utils.schemas.utils

URL: https://docs.axolotl.ai/docs/api/utils.schemas.utils.html

Contents:

  • utils.schemas.utils
  • Functions
    • handle_legacy_message_fields_logic
      • Parameters
      • Returns
      • Raises

Utilities for Axolotl Pydantic models

Handle backwards compatibility between legacy message field mapping and new property mapping system.

Previously, the config only supported mapping role and content fields via dedicated config options: - message_field_role: Mapped to the role field - message_field_content: Mapped to the content field

The new system uses message_property_mappings to support arbitrary field mappings: message_property_mappings: role: source_role_field content: source_content_field additional_field: source_field

Examples:

Example 1 (python):

utils.schemas.utils.handle_legacy_message_fields_logic(data)

prompt_strategies.alpaca_instruct

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.alpaca_instruct.html

Contents:

  • prompt_strategies.alpaca_instruct

prompt_strategies.alpaca_instruct

Module loading the AlpacaInstructPromptTokenizingStrategy class


utils.callbacks.lisa

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.lisa.html

Contents:

  • utils.callbacks.lisa

Adapted from https://github.com/OptimalScale/LMFlow/pull/701 for HF transformers & Axolotl Arxiv: https://arxiv.org/abs/2403.17919 License: Apache 2.0


models.mamba.modeling_mamba

URL: https://docs.axolotl.ai/docs/api/models.mamba.modeling_mamba.html

Contents:

  • models.mamba.modeling_mamba

models.mamba.modeling_mamba


prompt_strategies.metharme

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.metharme.html

Contents:

  • prompt_strategies.metharme
  • Classes
    • MetharmePromptTokenizingStrategy
    • MetharmePrompter

prompt_strategies.metharme

Module containing the MetharmenPromptTokenizingStrategy and MetharmePrompter class

Tokenizing strategy for the Metharme models

Prompter for the Metharme models.

Examples:

Example 1 (python):

prompt_strategies.metharme.MetharmePromptTokenizingStrategy(
    prompter,
    tokenizer,
    train_on_inputs=False,
    sequence_len=2048,
)

Example 2 (python):

prompt_strategies.metharme.MetharmePrompter(*args, **kwargs)

core.trainers.mamba

URL: https://docs.axolotl.ai/docs/api/core.trainers.mamba.html

Contents:

  • core.trainers.mamba
  • Classes
    • AxolotlMambaTrainer

Module for mamba trainer

Mamba specific trainer to handle loss calculation

Examples:

Example 1 (python):

core.trainers.mamba.AxolotlMambaTrainer(
    *_args,
    bench_data_collator=None,
    eval_data_collator=None,
    dataset_tags=None,
    **kwargs,
)

utils.ctx_managers.sequence_parallel

URL: https://docs.axolotl.ai/docs/api/utils.ctx_managers.sequence_parallel.html

Contents:

  • utils.ctx_managers.sequence_parallel
  • Classes
    • AllGatherWithGrad
      • Methods
        • backward
          • Parameters
          • Returns
        • forward
          • Parameters
          • Returns

utils.ctx_managers.sequence_parallel

Module for Axolotl trainer sequence parallelism manager and utilities

Custom autograd function for all-gather to preserve gradients.

Backward pass for all-gather operation.

Extracts the gradient slice corresponding to this ranks original input from the full gradient tensor.

Forward pass of all-gather of data with sequence dimension.

Context manager for sequence parallelism operations.

This class provides a context that will automatically apply sequence parallelism during model forward passes using a pre-forward hook, and gather outputs from across the sequence parallelism group using a post-forward hook.

Apply sequence parallelism slicing to a batch.

Special handling is implemented for integer logits_to_keep, which indicates to only keep the last N tokens in the sequence during generation.

Examples:

Example 1 (python):

utils.ctx_managers.sequence_parallel.AllGatherWithGrad()

Example 2 (python):

utils.ctx_managers.sequence_parallel.AllGatherWithGrad.backward(
    ctx,
    grad_output,
)

Example 3 (python):

utils.ctx_managers.sequence_parallel.AllGatherWithGrad.forward(
    ctx,
    input_tensor,
    group,
)

Example 4 (python):

utils.ctx_managers.sequence_parallel.SequenceParallelContextManager(
    models,
    context_parallel_size,
    gradient_accumulation_steps,
    ring_attn_func,
    heads_k_stride,
    gather_outputs,
    device_mesh=None,
)

utils.callbacks.qat

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.qat.html

Contents:

  • utils.callbacks.qat
  • Classes
    • QATCallback
  • Functions
    • toggle_fake_quant
      • Parameters

QAT Callback for HF Causal Trainer

Callback to toggle fake quantization for the model.

Toggle fake quantization for any fake quantized linear or embedding layers in the model.

Examples:

Example 1 (python):

utils.callbacks.qat.QATCallback(cfg)

Example 2 (python):

utils.callbacks.qat.toggle_fake_quant(mod, enable)

prompt_strategies.dpo.zephyr

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.zephyr.html

Contents:

  • prompt_strategies.dpo.zephyr

prompt_strategies.dpo.zephyr

DPO strategies for zephyr


kernels.utils

URL: https://docs.axolotl.ai/docs/api/kernels.utils.html

Contents:

  • kernels.utils

Utilities for axolotl.kernels submodules.


monkeypatch.multipack

URL: https://docs.axolotl.ai/docs/api/monkeypatch.multipack.html

Contents:

  • monkeypatch.multipack

monkeypatch.multipack

multipack patching for v2 of sample packing


cli.main

URL: https://docs.axolotl.ai/docs/api/cli.main.html

Contents:

  • cli.main
  • Functions
    • cli
    • evaluate
      • Parameters
    • fetch
      • Parameters
    • inference
      • Parameters
    • merge_lora

Click CLI definitions for various axolotl commands.

Axolotl CLI - Train and fine-tune large language models

Fetch example configs or other resources.

Available directories: - examples: Example configuration files - deepspeed_configs: DeepSpeed configuration files

Run inference with a trained model.

Merge trained LoRA adapters into a base model.

Merge sharded FSDP model weights.

Preprocess datasets before training.

Train or fine-tune a model.

Examples:

Example 1 (python):

cli.main.cli()

Example 2 (python):

cli.main.evaluate(ctx, config, launcher, **kwargs)

Example 3 (python):

cli.main.fetch(directory, dest)

Example 4 (python):

cli.main.inference(ctx, config, launcher, gradio, **kwargs)

core.trainers.mixins.optimizer

URL: https://docs.axolotl.ai/docs/api/core.trainers.mixins.optimizer.html

Contents:

  • core.trainers.mixins.optimizer
  • Classes
    • OptimizerInitMixin
    • OptimizerMixin

core.trainers.mixins.optimizer

Module for Axolotl trainer optimizer mixin

Mixin to handle common optimizer initialization logic for Trainers (mostly TRL) that do not accept optimizer_cls_and_kwargs as kwarg in constructor.

Mixin class for shared handling of building custom optimizers

Examples:

Example 1 (python):

core.trainers.mixins.optimizer.OptimizerInitMixin(*args, **kwargs)

Example 2 (python):

core.trainers.mixins.optimizer.OptimizerMixin()

integrations.kd.trainer

URL: https://docs.axolotl.ai/docs/api/integrations.kd.trainer.html

Contents:

  • integrations.kd.trainer
  • Classes
    • AxolotlKDTrainer
      • Methods
        • compute_loss

integrations.kd.trainer

Custom trainer subclass for Knowledge Distillation (KD)

How the loss is computed by Trainer. By default, all models return the loss in the first element.

Subclass and override for custom behavior.

Examples:

Example 1 (python):

integrations.kd.trainer.AxolotlKDTrainer(*args, **kwargs)

Example 2 (python):

integrations.kd.trainer.AxolotlKDTrainer.compute_loss(
    model,
    inputs,
    return_outputs=False,
    num_items_in_batch=None,
)

integrations.lm_eval.args

URL: https://docs.axolotl.ai/docs/api/integrations.lm_eval.args.html

Contents:

  • integrations.lm_eval.args
  • Classes
    • LMEvalArgs

integrations.lm_eval.args

Module for handling lm eval harness input arguments.

Input args for lm eval harness

Examples:

Example 1 (python):

integrations.lm_eval.args.LMEvalArgs()

integrations.cut_cross_entropy.args

URL: https://docs.axolotl.ai/docs/api/integrations.cut_cross_entropy.args.html

Contents:

  • integrations.cut_cross_entropy.args
  • Classes
    • CutCrossEntropyArgs

integrations.cut_cross_entropy.args

Module for handling Cut Cross Entropy input arguments.

Input args for Cut Cross Entropy.

Examples:

Example 1 (python):

integrations.cut_cross_entropy.args.CutCrossEntropyArgs()

monkeypatch.mistral_attn_hijack_flash

URL: https://docs.axolotl.ai/docs/api/monkeypatch.mistral_attn_hijack_flash.html

Contents:

  • monkeypatch.mistral_attn_hijack_flash

monkeypatch.mistral_attn_hijack_flash

Flash attention monkey patch for mistral model


loaders.constants

URL: https://docs.axolotl.ai/docs/api/loaders.constants.html

Contents:

  • loaders.constants

Shared constants for axolotl.loaders module


utils.bench

URL: https://docs.axolotl.ai/docs/api/utils.bench.html

Contents:

  • utils.bench
  • Functions
    • check_cuda_device

Benchmarking and measurement utilities

wraps a function and returns the default value instead of running the wrapped function if cuda isnt available or the device is auto :param default_value: :return:

Examples:

Example 1 (python):

utils.bench.check_cuda_device(default_value)

utils.trainer

URL: https://docs.axolotl.ai/docs/api/utils.trainer.html

Contents:

  • utils.trainer
  • Functions
    • add_pose_position_ids
    • add_position_ids
    • drop_long_seq
    • setup_trainer
      • Parameters
      • Returns

Module containing the Trainer class and related functions

use the PoSE technique to extend the context length by randomly skipping positions in the context. We only want to skip right before tokens in the split_on_token_ids list. We should attempt to randomly distribute the skips, but we dont need the final position_ids to be the full context_len. There may be multiple turns in the context, so we want to make sure we take into account the maximum possible number of skips remaining in each sample.

Handle both single-example and batched data. - single example: sample[input_ids] is a list[int] - batched data: sample[input_ids] is a list[list[int]]

Drop samples whose sequence length is either too long (> sequence_len) or too short (< min_sequence_len).

Works for both single-example (list[int]) or batched (list[list[int]]).

Helper method for instantiating and building a (causal or RLHF) trainer.

Examples:

Example 1 (python):

utils.trainer.add_pose_position_ids(
    sample,
    max_context_len=32768,
    split_on_token_ids=None,
    chunks=2,
)

Example 2 (python):

utils.trainer.add_position_ids(sample)

Example 3 (python):

utils.trainer.drop_long_seq(sample, sequence_len=2048, min_sequence_len=2)

Example 4 (python):

utils.trainer.setup_trainer(
    cfg,
    train_dataset,
    eval_dataset,
    model,
    tokenizer,
    processor,
    total_num_steps,
    model_ref=None,
    peft_config=None,
)

utils.schemas.config

URL: https://docs.axolotl.ai/docs/api/utils.schemas.config.html

Contents:

  • utils.schemas.config
  • Classes
    • AxolotlConfigWCapabilities
    • AxolotlInputConfig

Module with Pydantic models for configuration.

wrapper to valdiate GPU capabilities with the configured options

Wrapper of all config options.

Examples:

Example 1 (python):

utils.schemas.config.AxolotlConfigWCapabilities()

Example 2 (python):

utils.schemas.config.AxolotlInputConfig()

cli.args

URL: https://docs.axolotl.ai/docs/api/cli.args.html

Contents:

  • cli.args
  • Classes
    • EvaluateCliArgs
    • InferenceCliArgs
    • PreprocessCliArgs
    • QuantizeCliArgs
    • TrainerCliArgs
    • VllmServeCliArgs

Module for axolotl CLI command arguments.

Dataclass with CLI arguments for axolotl evaluate command.

Dataclass with CLI arguments for axolotl inference command.

Dataclass with CLI arguments for axolotl preprocess command.

Dataclass with CLI arguments for axolotl quantize command.

Dataclass with CLI arguments for axolotl train command.

Dataclass with CLI arguments for axolotl vllm-serve command.

Examples:

Example 1 (python):

cli.args.EvaluateCliArgs(
    debug=False,
    debug_text_only=False,
    debug_num_examples=0,
)

Example 2 (python):

cli.args.InferenceCliArgs(prompter=None)

Example 3 (python):

cli.args.PreprocessCliArgs(
    debug=False,
    debug_text_only=False,
    debug_num_examples=1,
    prompter=None,
    download=True,
    iterable=False,
)

Example 4 (python):

cli.args.QuantizeCliArgs(
    base_model=None,
    weight_dtype=None,
    activation_dtype=None,
    quantize_embedding=None,
    group_size=None,
    output_dir=None,
    hub_model_id=None,
)

common.architectures

URL: https://docs.axolotl.ai/docs/api/common.architectures.html

Contents:

  • common.architectures

Common architecture specific constants


cli.merge_sharded_fsdp_weights

URL: https://docs.axolotl.ai/docs/api/cli.merge_sharded_fsdp_weights.html

Contents:

  • cli.merge_sharded_fsdp_weights
  • Classes
    • BFloat16CastPlanner
  • Functions
    • do_cli
      • Parameters
    • merge_fsdp_weights
      • Parameters
      • Raises

cli.merge_sharded_fsdp_weights

CLI to merge sharded FSDP model checkpoints into a single combined checkpoint.

A custom planner to cast tensors to bfloat16 on the fly during loading.

Parses axolotl config, CLI args, and calls merge_fsdp_weights.

Merge the weights from sharded FSDP model checkpoints into a single combined checkpoint. Should be used if SHARDED_STATE_DICT was used for the model. Weights will be saved to {output_path}/model.safetensors if safe_serialization else pytorch_model.bin.

Note: this is a CPU-bound process.

Examples:

Example 1 (python):

cli.merge_sharded_fsdp_weights.BFloat16CastPlanner()

Example 2 (python):

cli.merge_sharded_fsdp_weights.do_cli(config=Path('examples/'), **kwargs)

Example 3 (python):

cli.merge_sharded_fsdp_weights.merge_fsdp_weights(
    checkpoint_dir,
    output_path,
    safe_serialization=False,
    remove_checkpoint_dir=False,
)

utils.data.streaming

URL: https://docs.axolotl.ai/docs/api/utils.data.streaming.html

Contents:

  • utils.data.streaming

Data handling specific to streaming datasets.


core.chat.format.chatml

URL: https://docs.axolotl.ai/docs/api/core.chat.format.chatml.html

Contents:

  • core.chat.format.chatml

core.chat.format.chatml

ChatML transformation functions for MessageContents


prompt_strategies.kto.chatml

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.kto.chatml.html

Contents:

  • prompt_strategies.kto.chatml
  • Functions
    • argilla_chat
    • intel
    • ultra

prompt_strategies.kto.chatml

KTO strategies for chatml

for argilla/kto-mix-15k conversations

For Intel Orca KTO ex: argilla/distilabel-intel-orca-kto

for ultrafeedback binarized conversations ex: argilla/ultrafeedback-binarized-preferences-cleaned-kto

Examples:

Example 1 (python):

prompt_strategies.kto.chatml.argilla_chat(cfg, **kwargs)

Example 2 (python):

prompt_strategies.kto.chatml.intel(cfg, **kwargs)

Example 3 (python):

prompt_strategies.kto.chatml.ultra(cfg, **kwargs)

utils.schemas.trl

URL: https://docs.axolotl.ai/docs/api/utils.schemas.trl.html

Contents:

  • utils.schemas.trl
  • Classes
    • TRLConfig

Pydantic models for TRL trainer configuration

Examples:

Example 1 (python):

utils.schemas.trl.TRLConfig()

monkeypatch.llama_attn_hijack_xformers

URL: https://docs.axolotl.ai/docs/api/monkeypatch.llama_attn_hijack_xformers.html

Contents:

  • monkeypatch.llama_attn_hijack_xformers

monkeypatch.llama_attn_hijack_xformers

Directly copied the code from https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/modules/llama_attn_hijack.py and made some adjustments


kernels.geglu

URL: https://docs.axolotl.ai/docs/api/kernels.geglu.html

Contents:

  • kernels.geglu
  • Functions
    • geglu_backward
      • Parameters
      • Returns
      • Note
    • geglu_forward
      • Parameters
      • Returns

Module for definition of GEGLU Triton kernels.

See “GLU Variants Improve Transformer” (https://arxiv.org/abs/2002.05202).

Credit to unsloth (https://unsloth.ai/) for inspiration for this implementation.

GEGLU backward pass using in-place operations.

This function modifies its input tensors in-place to store results.

Examples:

Example 1 (python):

kernels.geglu.geglu_backward(grad_output, gate, up)

Example 2 (python):

kernels.geglu.geglu_forward(gate, up)

utils.callbacks.profiler

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.profiler.html

Contents:

  • utils.callbacks.profiler
  • Classes
    • PytorchProfilerCallback

utils.callbacks.profiler

HF Trainer callback for creating pytorch profiling snapshots

PyTorch Profiler callback to create snapshots of GPU memory usage at specified steps.

Examples:

Example 1 (python):

utils.callbacks.profiler.PytorchProfilerCallback(
    steps_to_profile=5,
    profiler_steps_start=0,
)

kernels.lora

URL: https://docs.axolotl.ai/docs/api/kernels.lora.html

Contents:

  • kernels.lora
  • Classes
    • LoRA_MLP
      • Methods
        • backward
          • Parameters
          • Returns
        • forward
          • Parameters
          • Returns

Module for definition of Low-Rank Adaptation (LoRA) Triton kernels.

See “LoRA: Low-Rank Adaptation of Large Language Models” (https://arxiv.org/abs/2106.09685).

Credit to unsloth (https://unsloth.ai/) for inspiration for this implementation.

Optimized LoRA MLP implementation.

Performs backward pass computation for LoRA MLP.

Forward pass for LoRA MLP.

Optimized LoRA implementation for output projection.

Backward pass computing gradients for LoRA output projection.

Forward pass for output projection with LoRA.

Optimized LoRA QKV implementation with quantization support.

Implements efficient computation of query, key, value projections with LoRA, supporting quantization and memory optimization.

Backward pass computing gradients for LoRA QKV.

Forward pass computing Q, K, V projections with LoRA.

Applies LoRA to MLP layer with GEGLU activation.

Applies LoRA to MLP layer with SwiGLU activation.

Applies LoRA to output projection layer.

Applies LoRA to compute Query, Key, Value projections.

Gets LoRA parameters from a projection module.

Efficient fused matmul + LoRA computation.

Examples:

Example 1 (python):

kernels.lora.LoRA_MLP()

Example 2 (python):

kernels.lora.LoRA_MLP.backward(ctx, grad_output)

Example 3 (python):

kernels.lora.LoRA_MLP.forward(
    ctx,
    X,
    gate_weight,
    gate_bias,
    gate_quant,
    gate_A,
    gate_B,
    gate_scale,
    up_weight,
    up_bias,
    up_quant,
    up_A,
    up_B,
    up_scale,
    down_weight,
    down_bias,
    down_quant,
    down_A,
    down_B,
    down_scale,
    activation_fn,
    activation_fn_backward,
    inplace=True,
)

Example 4 (python):

kernels.lora.LoRA_O()

monkeypatch.trainer_fsdp_optim

URL: https://docs.axolotl.ai/docs/api/monkeypatch.trainer_fsdp_optim.html

Contents:

  • monkeypatch.trainer_fsdp_optim
  • Functions
    • patch_training_loop_for_fsdp

monkeypatch.trainer_fsdp_optim

fix for FSDP optimizer save in trainer w 4.47.0

monkeypatch for fixing the training loop for fsdp with optimizer save

Examples:

Example 1 (python):

monkeypatch.trainer_fsdp_optim.patch_training_loop_for_fsdp()

utils.schemas.multimodal

URL: https://docs.axolotl.ai/docs/api/utils.schemas.multimodal.html

Contents:

  • utils.schemas.multimodal
  • Classes
    • MultiModalConfig
      • Methods
        • convert_image_resize_algorithm

utils.schemas.multimodal

Pydantic models for multimodal-related configuration

Multi-modal configuration subset

Convert the image resize algorithm to a PIL.Image.Resampling enum.

Examples:

Example 1 (python):

utils.schemas.multimodal.MultiModalConfig()

Example 2 (python):

utils.schemas.multimodal.MultiModalConfig.convert_image_resize_algorithm(
    image_resize_algorithm,
)

prompt_strategies.dpo.llama3

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.llama3.html

Contents:

  • prompt_strategies.dpo.llama3
  • Functions
    • argilla_chat
    • icr
    • intel
    • ultra

prompt_strategies.dpo.llama3

DPO strategies for llama-3 chat template

for argilla/dpo-mix-7k conversations

chatml transforms for datasets with system, input, chosen, rejected ex. https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs

For Intel Orca DPO Pairs

for ultrafeedback binarized conversations

Examples:

Example 1 (python):

prompt_strategies.dpo.llama3.argilla_chat(cfg, **kwargs)

Example 2 (python):

prompt_strategies.dpo.llama3.icr(cfg, **kwargs)

Example 3 (python):

prompt_strategies.dpo.llama3.intel(cfg, **kwargs)

Example 4 (python):

prompt_strategies.dpo.llama3.ultra(cfg, **kwargs)

core.chat.format.shared

URL: https://docs.axolotl.ai/docs/api/core.chat.format.shared.html

Contents:

  • core.chat.format.shared

core.chat.format.shared

shared functions for format transforms


monkeypatch.llama_expand_mask

URL: https://docs.axolotl.ai/docs/api/monkeypatch.llama_expand_mask.html

Contents:

  • monkeypatch.llama_expand_mask

monkeypatch.llama_expand_mask

expands the binary attention mask per 3.2.2 of https://arxiv.org/pdf/2107.02027.pdf


core.chat.messages

URL: https://docs.axolotl.ai/docs/api/core.chat.messages.html

Contents:

  • core.chat.messages
  • Classes
    • ChatFormattedChats
    • Chats
    • MessageContentTypes
    • MessageContents
    • MessageRoles
    • Messages
    • PreferenceChats
    • SpecialToken

internal message representations of chat messages

Chat formatted chats with formatter and optional train on inputs

top level data structure for chat conversations

Message content types for text, image, audio, tool calls, and tool responses

Message contents with type, value, metadata, weight, newline, and end of contents

Message roles for the system, user, assistant, and tools

Messages with role, content, metadata, weight, and chat formatting

representation for preference data for chat

Special tokens for beginning of string and end of string

Tool with description, function, and parameters

Tool call contents with name, arguments, and optional id

Tool call function with name and arguments

Tool response contents with name, content, and optional id

Examples:

Example 1 (python):

core.chat.messages.ChatFormattedChats()

Example 2 (python):

core.chat.messages.Chats()

Example 3 (python):

core.chat.messages.MessageContentTypes()

Example 4 (python):

core.chat.messages.MessageContents()

core.datasets.transforms.chat_builder

URL: https://docs.axolotl.ai/docs/api/core.datasets.transforms.chat_builder.html

Contents:

  • core.datasets.transforms.chat_builder
  • Functions
    • chat_message_transform_builder
      • Parameters
      • Returns

core.datasets.transforms.chat_builder

This module contains a function that builds a transform that takes a row from the dataset and converts it to a Chat.

Builds a transform that takes a row from the dataset and converts it to a Chat

Examples:

Example 1 (python):

core.datasets.transforms.chat_builder.chat_message_transform_builder(
    train_on_inputs=False,
    conversations_field='messages',
    message_field_role=None,
    message_field_content=None,
    message_field_training=None,
)

utils.chat_templates

URL: https://docs.axolotl.ai/docs/api/utils.chat_templates.html

Contents:

  • utils.chat_templates

This module provides functionality for selecting chat templates based on user choices. These templates are used for formatting messages in a conversation.


core.trainers.dpo.trainer

URL: https://docs.axolotl.ai/docs/api/core.trainers.dpo.trainer.html

Contents:

  • core.trainers.dpo.trainer
  • Classes
    • AxolotlDPOTrainer
      • Methods
        • push_to_hub

core.trainers.dpo.trainer

DPO trainer for axolotl

Extend the base DPOTrainer for axolotl helpers.

Overwrite the push_to_hub method in order to force-add the tags when pushing the model on the Hub. Please refer to ~transformers.Trainer.push_to_hub for more details.

Examples:

Example 1 (python):

core.trainers.dpo.trainer.AxolotlDPOTrainer(*args, dataset_tags=None, **kwargs)

Example 2 (python):

core.trainers.dpo.trainer.AxolotlDPOTrainer.push_to_hub(*args, **kwargs)

monkeypatch.gradient_checkpointing.offload_disk

URL: https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_disk.html

Contents:

  • monkeypatch.gradient_checkpointing.offload_disk
  • Classes
    • Disco
      • Methods
        • backward
        • forward
        • get_instance
    • DiskOffloadManager
      • Methods
        • cleanup

monkeypatch.gradient_checkpointing.offload_disk

DISCO - DIsk-based Storage and Checkpointing with Optimized prefetching

Disco: DIsk-based Storage and Checkpointing with Optimized prefetching Advanced disk-based gradient checkpointer with prefetching.

Backward pass that loads activations from disk with prefetching

Forward pass that offloads activations to disk asynchronously

Get or create the offload manager

Manages offloaded tensors and handles prefetching in a separate thread. Includes synchronization to prevent race conditions.

Clean up all temp files and stop prefetch thread with proper synchronization

Clean up a specific tensor file after its been used

Load tensor from disk or prefetch cache with proper synchronization

Save tensor to disk asynchronously and return file path with thread-safe operations

Trigger prefetching of the next N tensors with proper synchronization

Wait for a tensor to be saved to disk

Examples:

Example 1 (python):

monkeypatch.gradient_checkpointing.offload_disk.Disco()

Example 2 (python):

monkeypatch.gradient_checkpointing.offload_disk.Disco.backward(
    ctx,
    *grad_outputs,
)

Example 3 (python):

monkeypatch.gradient_checkpointing.offload_disk.Disco.forward(
    ctx,
    forward_function,
    hidden_states,
    *args,
    prefetch_size=1,
    prefetch_to_gpu=True,
    save_workers=4,
)

Example 4 (python):

monkeypatch.gradient_checkpointing.offload_disk.Disco.get_instance(
    prefetch_size=1,
    prefetch_to_gpu=True,
    save_workers=4,
)

utils.samplers.multipack

URL: https://docs.axolotl.ai/docs/api/utils.samplers.multipack.html

Contents:

  • utils.samplers.multipack
  • Classes
    • MultipackBatchSampler
      • Methods
        • efficiency
        • gather_efficiency
          • Returns
        • gather_len_batches
        • generate_batches
          • Parameters

utils.samplers.multipack

Multipack Batch Sampler - An efficient batch sampler for packing variable-length sequences into fixed-capacity batches to optimize memory usage and training throughput.

Batch sampler class for efficient packing of variable-length sequences

This sampler packs sequences into fixed-capacity bins (batches) to maximize GPU memory utilization and training throughput by reducing padding.

It supports both parallel packing (using FFD algorithm) and sequential packing (preserving original sequence order).

Calculate the packing efficiency (ratio of tokens used to total token slots). Higher is better - 1.0 would mean perfect packing with no wasted space.

Gather and synchronize packing efficiency estimates across all distributed ranks.

Gather and synchronize batch counts across all distributed ranks. Returns the minimum number of batches available on any rank.

Generate packed batches for training.

Set the epoch number, used for reproducible shuffling across epochs

Sequential allocator that preserves example order.

First-fit-decreasing bin packing algorithm check.

Checks if sequences with the given lengths could fit in the specified number of bins.

Pack a group of sequences into bins using First-Fit Decreasing algorithm.

Pack sequences into bins using parallel processing.

Returns: List of bins, where each bin contains indices of sequences assigned to it.

Examples:

Example 1 (python):

utils.samplers.multipack.MultipackBatchSampler(
    sampler,
    batch_size,
    batch_max_len,
    lengths,
    packing_efficiency_estimate=1.0,
    drop_last=True,
    num_count_samples=4,
    sequential=False,
    group_size=100000,
    bin_size=200,
    num_processes=None,
    safe_mode=True,
    mp_start_method='fork',
    **kwargs,
)

Example 2 (python):

utils.samplers.multipack.MultipackBatchSampler.efficiency()

Example 3 (python):

utils.samplers.multipack.MultipackBatchSampler.gather_efficiency()

Example 4 (python):

utils.samplers.multipack.MultipackBatchSampler.gather_len_batches(num)

core.trainers.mixins.scheduler

URL: https://docs.axolotl.ai/docs/api/core.trainers.mixins.scheduler.html

Contents:

  • core.trainers.mixins.scheduler
  • Classes
    • SchedulerMixin
      • Methods
        • create_scheduler
          • Parameters

core.trainers.mixins.scheduler

Module for Axolotl trainer scheduler mixin

Mixin class for scheduler setup in CausalTrainer.

Set up the scheduler. The optimizer of the trainer must have been set up either before this method is called or passed as an argument.

Examples:

Example 1 (python):

core.trainers.mixins.scheduler.SchedulerMixin()

Example 2 (python):

core.trainers.mixins.scheduler.SchedulerMixin.create_scheduler(
    num_training_steps,
    optimizer=None,
)

utils.collators.batching

URL: https://docs.axolotl.ai/docs/api/utils.collators.batching.html

Contents:

  • utils.collators.batching
  • Classes
    • BatchSamplerDataCollatorForSeq2Seq
    • DataCollatorForSeq2Seq
      • Parameters
    • PretrainingBatchSamplerDataCollatorForSeq2Seq
    • V2BatchSamplerDataCollatorForSeq2Seq

utils.collators.batching

Data collators for axolotl to pad labels and position_ids for packed sequences

Collator for multipack specific to the using the BatchSampler

Data collator that will dynamically pad the inputs received, as well as the labels and position_ids

Collator for multipack specific to the using the BatchSampler

Collator for multipack specific to the using the BatchSampler

Examples:

Example 1 (python):

utils.collators.batching.BatchSamplerDataCollatorForSeq2Seq(
    tokenizer,
    model=None,
    padding=True,
    max_length=None,
    pad_to_multiple_of=None,
    label_pad_token_id=-100,
    position_pad_token_id=0,
    return_tensors='pt',
)

Example 2 (python):

utils.collators.batching.DataCollatorForSeq2Seq(
    tokenizer,
    model=None,
    padding=True,
    max_length=None,
    pad_to_multiple_of=None,
    label_pad_token_id=-100,
    position_pad_token_id=0,
    return_tensors='pt',
)

Example 3 (python):

utils.collators.batching.PretrainingBatchSamplerDataCollatorForSeq2Seq(
    *args,
    multipack_attn=True,
    **kwargs,
)

Example 4 (python):

utils.collators.batching.V2BatchSamplerDataCollatorForSeq2Seq(
    tokenizer,
    model=None,
    padding=True,
    max_length=None,
    pad_to_multiple_of=None,
    label_pad_token_id=-100,
    position_pad_token_id=0,
    return_tensors='pt',
    squash_position_ids=False,
)

prompt_strategies.orcamini

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.orcamini.html

Contents:

  • prompt_strategies.orcamini
  • Classes
    • OrcaMiniPrompter

prompt_strategies.orcamini

Prompt Strategy for finetuning Orca Mini (v2) models see also https://huggingface.co/psmathur/orca_mini_v2_7b for more information

Use dataset type: orcamini in config.yml to use this prompt style.

Compared to the alpaca_w_system.open_orca dataset type, this one specifies the system prompt with “### System:”.

Not suited/tested for multiple-turn conversations without further adjustments.

Adjusted Prompter for Orca Mini (v2) datasets

Examples:

Example 1 (python):

prompt_strategies.orcamini.OrcaMiniPrompter(
    prompt_style=PromptStyle.INSTRUCT.value,
)

prompt_strategies.dpo.chat_template

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.dpo.chat_template.html

Contents:

  • prompt_strategies.dpo.chat_template
  • Functions
    • argilla_chat
      • Parameters
      • Returns
      • Dataset format

prompt_strategies.dpo.chat_template

DPO prompt strategies for using tokenizer chat templates.

DPO chat template strategy for argilla-style datasets.

For argilla-style datasets where chosen/rejected contain full conversations instead of single response messages. Extracts the conversation history from the chosen field and formats both chosen/rejected responses using the configured chat template.

{ “chosen”: [ {“role”: “user”, “content”: “…”}, {“role”: “assistant”, “content”: “…”} ], “rejected”: [ {“role”: “user”, “content”: “…”}, {“role”: “assistant”, “content”: “…”} ] }

Examples:

Example 1 (python):

prompt_strategies.dpo.chat_template.argilla_chat(cfg, dataset_idx=0, **kwargs)

monkeypatch.relora

URL: https://docs.axolotl.ai/docs/api/monkeypatch.relora.html

Contents:

  • monkeypatch.relora
  • Classes
    • ReLoRACallback

Implements the ReLoRA training procedure from https://arxiv.org/abs/2307.05695, minus the initial full fine-tune.

Callback to merge LoRA weights into the base model and save full-weight checkpoints

Examples:

Example 1 (python):

monkeypatch.relora.ReLoRACallback(cfg)

monkeypatch.transformers_fa_utils

URL: https://docs.axolotl.ai/docs/api/monkeypatch.transformers_fa_utils.html

Contents:

  • monkeypatch.transformers_fa_utils
  • Functions
    • fixed_fa_peft_integration_check
      • Parameters

monkeypatch.transformers_fa_utils

see https://github.com/huggingface/transformers/pull/35834

PEFT usually casts the layer norms in float32 for training stability reasons therefore the input hidden states gets silently casted in float32. Hence, we need cast them back in float16 / bfloat16 just to be sure everything works as expected. This might slowdown training & inference so it is recommended to not cast the LayerNorms!

Examples:

Example 1 (python):

monkeypatch.transformers_fa_utils.fixed_fa_peft_integration_check(
    query,
    key,
    value,
    target_dtype=None,
    preferred_dtype=None,
)

utils.collators.mm_chat

URL: https://docs.axolotl.ai/docs/api/utils.collators.mm_chat.html

Contents:

  • utils.collators.mm_chat
  • Classes
    • MultiModalChatDataCollator

utils.collators.mm_chat

Collators for multi-modal chat messages and packing

Collator for multi-modal chat messages

Examples:

Example 1 (python):

utils.collators.mm_chat.MultiModalChatDataCollator(
    tokenizer,
    processing_strategy,
    packing=False,
    return_tensors='pt',
    padding=True,
    pad_to_multiple_of=None,
)

utils.lora

URL: https://docs.axolotl.ai/docs/api/utils.lora.html

Contents:

  • utils.lora
  • Functions
    • get_lora_merged_state_dict
      • Parameters
      • Returns

module to get the state dict of a merged lora model

Create and return a state_dict that has the LoRA deltas merged into the base models weights, without modifying model in place.

Examples:

Example 1 (python):

utils.lora.get_lora_merged_state_dict(model)

utils.model_shard_quant

URL: https://docs.axolotl.ai/docs/api/utils.model_shard_quant.html

Contents:

  • utils.model_shard_quant
  • Functions
    • load_and_quantize

utils.model_shard_quant

module to handle loading model on cpu/meta device for FSDP

Loads value tensor into submodule of module, optionally skipping skip_names and converting to dtype.

Quantizes Params4bit on device then places on “cpu” if to_cpu=True or “meta” if to_meta=True.

Examples:

Example 1 (python):

utils.model_shard_quant.load_and_quantize(
    module,
    name,
    value,
    device=None,
    dtype=None,
    skip_names=None,
    to_cpu=False,
    to_meta=False,
    verbose=False,
    quant_method='bnb',
)

monkeypatch.gradient_checkpointing.offload_cpu

URL: https://docs.axolotl.ai/docs/api/monkeypatch.gradient_checkpointing.offload_cpu.html

Contents:

  • monkeypatch.gradient_checkpointing.offload_cpu
  • Classes
    • CPU_Offloaded_Gradient_Checkpointer

monkeypatch.gradient_checkpointing.offload_cpu

CPU offloaded checkpointing

Saves VRAM by smartly offloading to RAM. Tiny hit to performance, since we mask the movement via non blocking calls.

Examples:

Example 1 (python):

monkeypatch.gradient_checkpointing.offload_cpu.CPU_Offloaded_Gradient_Checkpointer(
)

core.builders.base

URL: https://docs.axolotl.ai/docs/api/core.builders.base.html

Contents:

  • core.builders.base
  • Classes
    • TrainerBuilderBase
      • Methods
        • get_post_trainer_create_callbacks

Base class for trainer builder

Base class for trainer builder.

Callbacks added after the trainer is created, usually b/c these need access to the trainer

Examples:

Example 1 (python):

core.builders.base.TrainerBuilderBase(cfg, model, tokenizer, processor=None)

Example 2 (python):

core.builders.base.TrainerBuilderBase.get_post_trainer_create_callbacks(trainer)

core.builders.rl

URL: https://docs.axolotl.ai/docs/api/core.builders.rl.html

Contents:

  • core.builders.rl
  • Classes
    • HFRLTrainerBuilder

Builder for RLHF trainers

Trainer factory class for TRL-based RLHF trainers (e.g. DPO)

Examples:

Example 1 (python):

core.builders.rl.HFRLTrainerBuilder(cfg, model, tokenizer, processor=None)

utils.schemas.integrations

URL: https://docs.axolotl.ai/docs/api/utils.schemas.integrations.html

Contents:

  • utils.schemas.integrations
  • Classes
    • CometConfig
    • GradioConfig
    • LISAConfig
    • MLFlowConfig
    • OpenTelemetryConfig
    • RayConfig
    • WandbConfig

utils.schemas.integrations

Pydantic models for Axolotl integrations

Comet configuration subset

Gradio configuration subset

LISA configuration subset

MLFlow configuration subset

OpenTelemetry configuration subset

Ray launcher configuration subset

Wandb configuration subset

Examples:

Example 1 (python):

utils.schemas.integrations.CometConfig()

Example 2 (python):

utils.schemas.integrations.GradioConfig()

Example 3 (python):

utils.schemas.integrations.LISAConfig()

Example 4 (python):

utils.schemas.integrations.MLFlowConfig()

utils.data.sft

URL: https://docs.axolotl.ai/docs/api/utils.data.sft.html

Contents:

  • utils.data.sft
  • Functions
    • prepare_datasets
      • Parameters
      • Returns

Data handling specific to SFT.

Prepare training and evaluation datasets based on configuration.

Examples:

Example 1 (python):

utils.data.sft.prepare_datasets(cfg, tokenizer, processor=None)

integrations.liger.args

URL: https://docs.axolotl.ai/docs/api/integrations.liger.args.html

Contents:

  • integrations.liger.args
  • Classes
    • LigerArgs

integrations.liger.args

Module for handling LIGER input arguments.

Input args for LIGER.

Examples:

Example 1 (python):

integrations.liger.args.LigerArgs()

monkeypatch.mixtral

URL: https://docs.axolotl.ai/docs/api/monkeypatch.mixtral.html

Contents:

  • monkeypatch.mixtral

Patches to support multipack for mixtral


cli.preprocess

URL: https://docs.axolotl.ai/docs/api/cli.preprocess.html

Contents:

  • cli.preprocess
  • Functions
    • do_cli
      • Parameters
    • do_preprocess
      • Parameters

CLI to run preprocessing of a dataset.

Parses axolotl config, CLI args, and calls do_preprocess.

Preprocesses dataset specified in axolotl config.

Examples:

Example 1 (python):

cli.preprocess.do_cli(config=Path('examples/'), **kwargs)

Example 2 (python):

cli.preprocess.do_preprocess(cfg, cli_args)

prompt_strategies.kto.llama3

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.kto.llama3.html

Contents:

  • prompt_strategies.kto.llama3
  • Functions
    • argilla_chat
    • intel
    • ultra

prompt_strategies.kto.llama3

KTO strategies for llama-3 chat template

for argilla/kto-mix-15k conversations

For Intel Orca KTO ex: argilla/distilabel-intel-orca-kto

for ultrafeedback binarized conversations ex: argilla/ultrafeedback-binarized-preferences-cleaned-kto

Examples:

Example 1 (python):

prompt_strategies.kto.llama3.argilla_chat(cfg, **kwargs)

Example 2 (python):

prompt_strategies.kto.llama3.intel(cfg, **kwargs)

Example 3 (python):

prompt_strategies.kto.llama3.ultra(cfg, **kwargs)

prompt_strategies.orpo.chat_template

URL: https://docs.axolotl.ai/docs/api/prompt_strategies.orpo.chat_template.html

Contents:

  • prompt_strategies.orpo.chat_template
  • Classes
    • Message
    • MessageList
    • ORPODatasetParsingStrategy
      • Methods
        • get_chosen_conversation_thread
        • get_prompt
        • get_rejected_conversation_thread
    • ORPOPrompter

prompt_strategies.orpo.chat_template

chatml prompt tokenization strategy for ORPO

Strategy to parse chosen rejected dataset into messagelist

Dataset structure mappings

Map the data to extract everything up to the last turn

Dataset structure mappings

Single Turn prompter for ORPO

rejected_input_ids input_ids rejected_attention_mask attention_mask rejected_labels labels

chatml transforms for datasets with system, input, chosen, rejected

Examples:

Example 1 (python):

prompt_strategies.orpo.chat_template.Message()

Example 2 (python):

prompt_strategies.orpo.chat_template.MessageList()

Example 3 (python):

prompt_strategies.orpo.chat_template.ORPODatasetParsingStrategy()

Example 4 (python):

prompt_strategies.orpo.chat_template.ORPODatasetParsingStrategy.get_chosen_conversation_thread(
    prompt,
)

loaders.processor

URL: https://docs.axolotl.ai/docs/api/loaders.processor.html

Contents:

  • loaders.processor

Processor loading functionality for multi-modal models


utils.callbacks.comet_

URL: https://docs.axolotl.ai/docs/api/utils.callbacks.comet_.html

Contents:

  • utils.callbacks.comet_
  • Classes
    • SaveAxolotlConfigtoCometCallback

utils.callbacks.comet_

Comet module for trainer callbacks

Callback to save axolotl config to comet

Examples:

Example 1 (python):

utils.callbacks.comet_.SaveAxolotlConfigtoCometCallback(axolotl_config_path)