Fine-tuning

Finetuning Quickstart – learn how to fine-tune an LLM in 5 mins.

1. Register for an account

First, register for an account to get an API key. New accounts come with $1 to get started.

Once you've registered, set your account's API key to an environment variable named TOGETHER_API_KEY:

export TOGETHER_API_KEY=xxxxx

2. Install your preferred library

Together provides an official library for Python:

pip install together --upgrade
npm install together-ai

3. Fine-tuning Dataset

We will use a subset of the CoQA conversational dataset, download the formatted dataset here.

This is what one row/sample from the CoQA dataset looks like in conversation format:

{'messages':  
    [  
     {
       'content': 'Read the story and extract answers for the questions.\nStory: An incandescent light bulb, incandescent lamp or incandescent ...',
       'role': 'system'
     },  
     {
       'content': 'What is the energy source for an incandescent bulb?', 
       'role': 'user'
     },  
     {
       'content': 'a wire filament', 
       'role': 'assistant'
     },  
     {
       'content': 'Is it hot?', 
       'role': 'user'
     },  
     {'content': 'yes', 
      'role': 'assistant'
     },  
     ...  
    ]  
 }

4. Check and Upload Dataset

To upload your data, use the CLI or our Python library (our TypeScript library currently doesn't support file uploads):

together files check "small_coqa.jsonl" 

together files upload "small_coqa.jsonl"
import os
from together import Together

client = Together(api_key=os.environ.get("TOGETHER_API_KEY"))

file_resp = client.files.upload(file="small_coqa.jsonl", check=True)

print(file_resp.model_dump())
import Together from 'together-ai';
import fs from 'fs';

const client = new Together({
  apiKey: process.env.TOGETHER_API_KEY
});

const fileContent = fs.readFileSync('small_coqa.jsonl');
const fileResponse = await client.files.upload({
  file: fileContent,
  check: true
});
    
console.log(fileResponse);

You'll see the following output once the upload finishes:

{
  id='file-629e58b4-ff73-438c-b2cc-f69542b27980', 
  object=<ObjectType.File: 'file'>, 
  created_at=1732573871, 
  type=None, 
  purpose=<FilePurpose.FineTune: 'fine-tune'>, 
  filename='small_coqa.jsonl', 
  bytes=0, 
  line_count=0, 
  processed=False, 
  FileType='jsonl'
}

You'll be using your file's ID (the string that begins with "file-") to start your fine-tuning job, so store it somewhere before moving on.

You're now ready to kick off your first fine-tuning job!

5. Starting a fine-tuning job

We support both LoRA and full fine-tuning – see how to start a finetuning job with either method below.

together fine-tuning create \
  --training-file "file-629e58b4-ff73-438c-b2cc-f69542b27980" \
  --model "meta-llama/Meta-Llama-3.1-8B-Instruct-Reference" \
  --lora
# Trigger fine-tuning job

response = client.fine_tuning.create(
  training_file = file_resp.id,
  model = 'meta-llama/Meta-Llama-3.1-8B-Instruct-Reference',
  lora = True,
)

print(response)
const response = await client.fineTuning.create({
  training_file: fileResponse.id,
  model: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Reference',
  lora: true
});
    
console.log(response);

You can specify many more fine-tuning parameters to customize your job. See the full list of hyperparameters and their definitions here.

The response object will have all the details of your job, including its ID and a status key that starts out as "pending":

{
  id='ft-66592697-0a37-44d1-b6ea-9908d1c81fbd', 
  training_file='file-63b9f097-e582-4d2e-941e-4b541aa7e328', 
  validation_file='', 
  model='meta-llama/Meta-Llama-3.1-8B-Instruct-Reference', 
  output_name='zainhas/Meta-Llama-3.1-8B-Instruct-Reference-30b975fd', 
... 
  status=<FinetuneJobStatus.STATUS_PENDING: 'pending'>
}

6. Monitoring a fine-tuning job's progress

After you started your job, visit your jobs dashboard. You should see your new job!

You can also pass your Job ID to retrieve to get the latest details about your job directly from your code:

together fine-tuning retrieve "ft-66592697-0a37-44d1-b6ea-9908d1c81fbd"
response = client.fine_tuning.retrieve('ft-66592697-0a37-44d1-b6ea-9908d1c81fbd')

print(response.status) # STATUS_UPLOADING

Your fine-tuning job will go through several phases, including Pending, Queued, Running, Uploading, and Completed.

7. Deploying your fine-tuned model

Once your fine-tune job completes, you should see your new model in your models dashboard:

To use your model, you can either host it on Together AI, click View Deploy for an hourly usage fee, or download your model checkpoint and run it locally.

For more details read the detailed walkthropugh How-to: Fine-tuning .