Skip to main content
The following models are available to use with our fine-tuning API. Get started with fine-tuning a model! Note: This list is different from the models that support Serverless LoRA inference, which allows you to perform LoRA fine-tuning and run inference immediately. See the LoRA inference page for the list of supported base models for serverless LoRA. Important: When uploading LoRA adapters for serverless inference, you must use base models from the serverless LoRA list, not the fine-tuning models list. Using an incompatible base model (such as Turbo variants) will result in a “No lora_model specified” error during upload. For example, use meta-llama/Meta-Llama-3.1-8B-Instruct-Reference instead of meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo for serverless LoRA adapters.
  • Training Precision Type indicates the precision type used during training for each model.
    • AMP (Automated Mixed Precision): AMP allows the training speed to be faster with less memory usage while preserving convergence behavior compared to using float32. Learn more about AMP in this PyTorch blog.
    • bf16 (bfloat 16): This uses bf16 for all weights. Some large models on our platform use full bf16 training for better memory usage and training speed.
  • For batch sizes of 1, Gradient accumulation 8 is used, so effectively you will get batch size 8 (iteration time is slower).
  • Long-context fine-tuning of Llama 3.1 (8B) Reference, Llama 3.1 (70B) Reference, Llama 3.1 Instruct (70B) Reference for context sizes of 32K-131K is only supported using the LoRA method.
  • For Llama 3.1 (405B) Fine-tuning, please contact us.
Request a model

LoRA Fine-tuning

OrganizationModel NameModel String for APIContext Length (SFT)Context Length (DPO)Max Batch Size (SFT)Max Batch Size (DPO)Min Batch Size
OpenAIgpt-oss-20bopenai/gpt-oss-20b163848192888
OpenAIgpt-oss-120bopenai/gpt-oss-120b163848192161616
DeepSeekDeepSeek-R1-0528deepseek-ai/DeepSeek-R1-052813107232768112
DeepSeekDeepSeek-R1deepseek-ai/DeepSeek-R113107232768112
DeepSeekDeepSeek-V3.1deepseek-ai/DeepSeek-V3.113107232768112
DeepSeekDeepSeek-V3-0324deepseek-ai/DeepSeek-V3-032413107232768112
DeepSeekDeepSeek-V3deepseek-ai/DeepSeek-V313107232768112
DeepSeekDeepSeek-V3.1-Basedeepseek-ai/DeepSeek-V3.1-Base13107232768112
DeepSeekDeepSeek-V3-Basedeepseek-ai/DeepSeek-V3-Base13107232768112
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B2457612288888
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B-32k3276816384111
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B-131k13107216384111
DeepSeekDeepSeek-R1-Distill-Qwen-14Bdeepseek-ai/DeepSeek-R1-Distill-Qwen-14B6553649152888
DeepSeekDeepSeek-R1-Distill-Qwen-1.5Bdeepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B131072131072888
MetaLlama-4-Scout-17B-16Emeta-llama/Llama-4-Scout-17B-16E1638412288888
MetaLlama-4-Scout-17B-16E-Instructmeta-llama/Llama-4-Scout-17B-16E-Instruct1638412288888
MetaLlama-4-Maverick-17B-128Emeta-llama/Llama-4-Maverick-17B-128E1638424576161616
MetaLlama-4-Maverick-17B-128E-Instructmeta-llama/Llama-4-Maverick-17B-128E-Instruct1638424576161616
Googlegemma-3-270mgoogle/gemma-3-270m32768327681281288
Googlegemma-3-270m-itgoogle/gemma-3-270m-it32768327681281288
Googlegemma-3-1b-itgoogle/gemma-3-1b-it327683276832328
Googlegemma-3-1b-ptgoogle/gemma-3-1b-pt327683276832328
Googlegemma-3-4b-itgoogle/gemma-3-4b-it13107265536888
Googlegemma-3-4b-ptgoogle/gemma-3-4b-pt13107265536888
Googlegemma-3-12b-itgoogle/gemma-3-12b-it1638449152888
Googlegemma-3-12b-ptgoogle/gemma-3-12b-pt6553649152888
Googlegemma-3-27b-itgoogle/gemma-3-27b-it4915224576888
Googlegemma-3-27b-ptgoogle/gemma-3-27b-pt4915224576888
QwenQwen3-Next-80B-A3B-InstructQwen/Qwen3-Next-80B-A3B-Instruct6553616384888
QwenQwen3-Next-80B-A3B-ThinkingQwen/Qwen3-Next-80B-A3B-Thinking6553616384888
QwenQwen3-0.6BQwen/Qwen3-0.6B327684096064648
QwenQwen3-0.6B-BaseQwen/Qwen3-0.6B-Base327683276864648
QwenQwen3-1.7BQwen/Qwen3-1.7B327684096032328
QwenQwen3-1.7B-BaseQwen/Qwen3-1.7B-Base327683276832328
QwenQwen3-4BQwen/Qwen3-4B327684096016168
QwenQwen3-4B-BaseQwen/Qwen3-4B-Base327683276816168
QwenQwen3-8BQwen/Qwen3-8B3276840960888
QwenQwen3-8B-BaseQwen/Qwen3-8B-Base327683276816168
QwenQwen3-14BQwen/Qwen3-14B3276840960888
QwenQwen3-14B-BaseQwen/Qwen3-14B-Base3276840960888
QwenQwen3-32BQwen/Qwen3-32B2457624576888
QwenQwen3-30B-A3B-BaseQwen/Qwen3-30B-A3B-Base81923276816168
QwenQwen3-30B-A3BQwen/Qwen3-30B-A3B81923276816168
QwenQwen3-30B-A3B-Instruct-2507Qwen/Qwen3-30B-A3B-Instruct-250781923276816168
QwenQwen3-235B-A22BQwen/Qwen3-235B-A22B3276824576118
QwenQwen3-235B-A22B-Instruct-2507Qwen/Qwen3-235B-A22B-Instruct-25073276824576118
QwenQwen3-Coder-30B-A3B-InstructQwen/Qwen3-Coder-30B-A3B-Instruct8192819216168
QwenQwen3-Coder-480B-A35B-InstructQwen/Qwen3-Coder-480B-A35B-Instruct13107232768112
MetaLlama-3.3-70B-Instruct-Referencemeta-llama/Llama-3.3-70B-Instruct-Reference245768192888
MetaLlama-3.3-70B-32k-Instruct-Referencemeta-llama/Llama-3.3-70B-32k-Instruct-Reference3276865536111
MetaLlama-3.3-70B-131k-Instruct-Referencemeta-llama/Llama-3.3-70B-131k-Instruct-Reference13107265536111
MetaLlama-3.2-3B-Instructmeta-llama/Llama-3.2-3B-Instruct13107265536888
MetaLlama-3.2-3Bmeta-llama/Llama-3.2-3B13107265536888
MetaLlama-3.2-1B-Instructmeta-llama/Llama-3.2-1B-Instruct131072131072888
MetaLlama-3.2-1Bmeta-llama/Llama-3.2-1B131072131072888
MetaMeta-Llama-3.1-8B-Instruct-Referencemeta-llama/Meta-Llama-3.1-8B-Instruct-Reference13107265536888
MetaMeta-Llama-3.1-8B-131k-Instruct-Referencemeta-llama/Meta-Llama-3.1-8B-131k-Instruct-Reference131072131072441
MetaMeta-Llama-3.1-8B-Referencemeta-llama/Meta-Llama-3.1-8B-Reference13107265536888
MetaMeta-Llama-3.1-8B-131k-Referencemeta-llama/Meta-Llama-3.1-8B-131k-Reference131072131072441
MetaMeta-Llama-3.1-70B-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-Instruct-Reference2457612288888
MetaMeta-Llama-3.1-70B-32k-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-32k-Instruct-Reference3276832768111
MetaMeta-Llama-3.1-70B-131k-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-131k-Instruct-Reference13107265536111
MetaMeta-Llama-3.1-70B-Referencemeta-llama/Meta-Llama-3.1-70B-Reference2457612288888
MetaMeta-Llama-3.1-70B-32k-Referencemeta-llama/Meta-Llama-3.1-70B-32k-Reference3276832768111
MetaMeta-Llama-3.1-70B-131k-Referencemeta-llama/Meta-Llama-3.1-70B-131k-Reference13107265536111
MetaMeta-Llama-3-8B-Instructmeta-llama/Meta-Llama-3-8B-Instruct8192819264648
MetaMeta-Llama-3-8Bmeta-llama/Meta-Llama-3-8B8192819264648
MetaMeta-Llama-3-70B-Instructmeta-llama/Meta-Llama-3-70B-Instruct81928192888
QwenQwen2.5-72B-InstructQwen/Qwen2.5-72B-Instruct3276812288888
QwenQwen2.5-72BQwen/Qwen2.5-72B2457612288888
QwenQwen2.5-32B-InstructQwen/Qwen2.5-32B-Instruct3276832768888
QwenQwen2.5-32BQwen/Qwen2.5-32B4915232768888
QwenQwen2.5-14B-InstructQwen/Qwen2.5-14B-Instruct3276832768888
QwenQwen2.5-14BQwen/Qwen2.5-14B6553649152888
QwenQwen2.5-7B-InstructQwen/Qwen2.5-7B-Instruct327683276816168
QwenQwen2.5-7BQwen/Qwen2.5-7B13107265536888
QwenQwen2.5-3B-InstructQwen/Qwen2.5-3B-Instruct327683276832328
QwenQwen2.5-3BQwen/Qwen2.5-3B327683276832328
QwenQwen2.5-1.5B-InstructQwen/Qwen2.5-1.5B-Instruct327683276832328
QwenQwen2.5-1.5BQwen/Qwen2.5-1.5B32768131072888
QwenQwen2-72B-InstructQwen/Qwen2-72B-Instruct3276812288161616
QwenQwen2-72BQwen/Qwen2-72B3276812288161616
QwenQwen2-7B-InstructQwen/Qwen2-7B-Instruct3276832768888
QwenQwen2-7BQwen/Qwen2-7B13107224576888
QwenQwen2-1.5B-InstructQwen/Qwen2-1.5B-Instruct327683276832328
QwenQwen2-1.5BQwen/Qwen2-1.5B131072131072888
MistralMixtral-8x7B-Instruct-v0.1mistralai/Mixtral-8x7B-Instruct-v0.13276832768888
MistralMixtral-8x7B-v0.1mistralai/Mixtral-8x7B-v0.13276832768888
MistralMistral-7B-Instruct-v0.2mistralai/Mistral-7B-Instruct-v0.2327683276816168
MistralMistral-7B-v0.1mistralai/Mistral-7B-v0.1327683276816168
TekniumOpenHermes-2p5-Mistral-7Bteknium/OpenHermes-2p5-Mistral-7B327683276816168
MetaCodeLlama-7b-hfcodellama/CodeLlama-7b-hf163841638416168
Togetherllama-2-7b-chattogethercomputer/llama-2-7b-chat4096409664648

LoRA Long-context Fine-tuning

OrganizationModel NameModel String for APIContext Length (SFT)Context Length (DPO)Max Batch Size (SFT)Max Batch Size (DPO)Min Batch Size
DeepSeekDeepSeek-R1-0528deepseek-ai/DeepSeek-R1-052813107232768112
DeepSeekDeepSeek-R1deepseek-ai/DeepSeek-R113107232768112
DeepSeekDeepSeek-V3.1deepseek-ai/DeepSeek-V3.113107232768112
DeepSeekDeepSeek-V3-0324deepseek-ai/DeepSeek-V3-032413107232768112
DeepSeekDeepSeek-V3deepseek-ai/DeepSeek-V313107232768112
DeepSeekDeepSeek-V3.1-Basedeepseek-ai/DeepSeek-V3.1-Base13107232768112
DeepSeekDeepSeek-V3-Basedeepseek-ai/DeepSeek-V3-Base13107232768112
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B-32k3276816384111
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B-131k13107216384111
QwenQwen3-235B-A22BQwen/Qwen3-235B-A22B3276824576118
QwenQwen3-235B-A22B-Instruct-2507Qwen/Qwen3-235B-A22B-Instruct-25073276824576118
QwenQwen3-Coder-480B-A35B-InstructQwen/Qwen3-Coder-480B-A35B-Instruct13107232768112
MetaLlama-3.3-70B-32k-Instruct-Referencemeta-llama/Llama-3.3-70B-32k-Instruct-Reference3276865536111
MetaLlama-3.3-70B-131k-Instruct-Referencemeta-llama/Llama-3.3-70B-131k-Instruct-Reference13107265536111
MetaMeta-Llama-3.1-8B-131k-Instruct-Referencemeta-llama/Meta-Llama-3.1-8B-131k-Instruct-Reference131072131072441
MetaMeta-Llama-3.1-8B-131k-Referencemeta-llama/Meta-Llama-3.1-8B-131k-Reference131072131072441
MetaMeta-Llama-3.1-70B-32k-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-32k-Instruct-Reference3276832768111
MetaMeta-Llama-3.1-70B-131k-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-131k-Instruct-Reference13107265536111
MetaMeta-Llama-3.1-70B-32k-Referencemeta-llama/Meta-Llama-3.1-70B-32k-Reference3276832768111
MetaMeta-Llama-3.1-70B-131k-Referencemeta-llama/Meta-Llama-3.1-70B-131k-Reference13107265536111

Full Fine-tuning

OrganizationModel NameModel String for APIContext Length (SFT)Context Length (DPO)Max Batch Size (SFT)Max Batch Size (DPO)Min Batch Size
DeepSeekDeepSeek-R1-Distill-Llama-70Bdeepseek-ai/DeepSeek-R1-Distill-Llama-70B2457612288323232
DeepSeekDeepSeek-R1-Distill-Qwen-14Bdeepseek-ai/DeepSeek-R1-Distill-Qwen-14B6553649152888
DeepSeekDeepSeek-R1-Distill-Qwen-1.5Bdeepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B131072131072888
Googlegemma-3-270mgoogle/gemma-3-270m32768327681281288
Googlegemma-3-270m-itgoogle/gemma-3-270m-it32768327681281288
Googlegemma-3-1b-itgoogle/gemma-3-1b-it327683276864648
Googlegemma-3-1b-ptgoogle/gemma-3-1b-pt327683276864648
Googlegemma-3-4b-itgoogle/gemma-3-4b-it13107265536888
Googlegemma-3-4b-ptgoogle/gemma-3-4b-pt13107265536888
Googlegemma-3-12b-itgoogle/gemma-3-12b-it1638449152888
Googlegemma-3-12b-ptgoogle/gemma-3-12b-pt6553649152888
Googlegemma-3-27b-itgoogle/gemma-3-27b-it4915224576161616
Googlegemma-3-27b-ptgoogle/gemma-3-27b-pt4915224576161616
QwenQwen3-0.6BQwen/Qwen3-0.6B327684096064648
QwenQwen3-0.6B-BaseQwen/Qwen3-0.6B-Base327683276864648
QwenQwen3-1.7BQwen/Qwen3-1.7B327684096032328
QwenQwen3-1.7B-BaseQwen/Qwen3-1.7B-Base327683276832328
QwenQwen3-4BQwen/Qwen3-4B327684096016168
QwenQwen3-4B-BaseQwen/Qwen3-4B-Base327683276816168
QwenQwen3-8BQwen/Qwen3-8B3276840960888
QwenQwen3-8B-BaseQwen/Qwen3-8B-Base327683276816168
QwenQwen3-14BQwen/Qwen3-14B3276840960888
QwenQwen3-14B-BaseQwen/Qwen3-14B-Base3276840960888
QwenQwen3-32BQwen/Qwen3-32B2457624576161616
QwenQwen3-30B-A3B-BaseQwen/Qwen3-30B-A3B-Base819232768888
QwenQwen3-30B-A3BQwen/Qwen3-30B-A3B819232768888
QwenQwen3-30B-A3B-Instruct-2507Qwen/Qwen3-30B-A3B-Instruct-2507819232768888
QwenQwen3-Coder-30B-A3B-InstructQwen/Qwen3-Coder-30B-A3B-Instruct81928192888
MetaLlama-3.3-70B-Instruct-Referencemeta-llama/Llama-3.3-70B-Instruct-Reference245768192323232
MetaLlama-3.2-3B-Instructmeta-llama/Llama-3.2-3B-Instruct13107265536888
MetaLlama-3.2-3Bmeta-llama/Llama-3.2-3B13107265536888
MetaLlama-3.2-1B-Instructmeta-llama/Llama-3.2-1B-Instruct131072131072888
MetaLlama-3.2-1Bmeta-llama/Llama-3.2-1B131072131072888
MetaMeta-Llama-3.1-8B-Instruct-Referencemeta-llama/Meta-Llama-3.1-8B-Instruct-Reference13107265536888
MetaMeta-Llama-3.1-8B-Referencemeta-llama/Meta-Llama-3.1-8B-Reference13107265536888
MetaMeta-Llama-3.1-70B-Instruct-Referencemeta-llama/Meta-Llama-3.1-70B-Instruct-Reference2457612288323232
MetaMeta-Llama-3.1-70B-Referencemeta-llama/Meta-Llama-3.1-70B-Reference2457612288323232
MetaMeta-Llama-3-8B-Instructmeta-llama/Meta-Llama-3-8B-Instruct8192819264648
MetaMeta-Llama-3-8Bmeta-llama/Meta-Llama-3-8B8192819264648
MetaMeta-Llama-3-70B-Instructmeta-llama/Meta-Llama-3-70B-Instruct81928192323232
QwenQwen2-7B-InstructQwen/Qwen2-7B-Instruct3276832768888
QwenQwen2-7BQwen/Qwen2-7B13107224576888
QwenQwen2-1.5B-InstructQwen/Qwen2-1.5B-Instruct327683276832328
QwenQwen2-1.5BQwen/Qwen2-1.5B131072131072888
MistralMixtral-8x7B-Instruct-v0.1mistralai/Mixtral-8x7B-Instruct-v0.13276832768161616
MistralMixtral-8x7B-v0.1mistralai/Mixtral-8x7B-v0.13276832768161616
MistralMistral-7B-Instruct-v0.2mistralai/Mistral-7B-Instruct-v0.2327683276816168
MistralMistral-7B-v0.1mistralai/Mistral-7B-v0.1327683276816168
TekniumOpenHermes-2p5-Mistral-7Bteknium/OpenHermes-2p5-Mistral-7B327683276816168
MetaCodeLlama-7b-hfcodellama/CodeLlama-7b-hf163841638416168
Togetherllama-2-7b-chattogethercomputer/llama-2-7b-chat4096409664648