Jump to Content
Documentation
API Reference
Web Playground
Log In
Documentation
Web Playground
Log In
Moon (Dark Mode)
Sun (Light Mode)
Documentation
API Reference
Search
Getting Started
Introduction
Quickstart
OpenAI compatibility
DeepSeek-R1 Quickstart
Reasoning Models Guide
Prompting DeepSeek-R1
DeepSeek FAQs
Inference
Serverless models
Dedicated models
Dedicated Inference
Uploading a fine-tuned model
Examples
Together Cookbooks
Example Apps
Capabilities
Chat
JSON Mode
Function calling
Logprobs
Integrations
Images
Quickstart: Flux Tools Models
Quickstart: Flux LoRA Inference
Vision
Text-to-Speech
Code/Language
Rerank
QuickStart: LlamaRank
Embeddings
RAG Integrations
Serverless LoRA Inference
CodeSandbox SDK (Code execution)
Agent Workflows
Sequential Workflow
Parallel Workflow
Conditional Workflow
Iterative Workflow
Training
Fine-tuning
How-to: Fine-tuning
Data preparation
Fine-tuning Models
GPU Clusters
Cluster user management
Cluster storage
Slurm management system
Guides
Quickstart: Retrieval Augmented Generation (RAG)
Quickstart: Next.js
Quickstart: Using Hugging Face Inference with Together
Quickstart: Using Vercel's AI SDK with Together AI
Together Mixture-Of-Agents (MoA)
Search and RAG
How to Improve Search with Rerankers
How to build an AI search engine (OSS Perplexity Clone)
Building a RAG Workflow
How to Implement Contextual RAG from Anthropic
Apps
How to build a Claude Artifacts Clone with Llama 3.1 405B
Building an AI data analyst
Fine-tuning Llama-3 to get 90% of GPT-4’s performance
How to build a real-time image generator with Flux and Together AI
How to build an Open Source NotebookLM: PDF to Podcast
How to build an Interactive AI Tutor with Llama 3.1
frequently asked questions
Deployment Options
Rate limits
Error codes
Deploy dedicated endpoints in the web
Priority Support
Create tickets in Slack
Customer Ticket Portal
Deprecations
Playground
Inference FAQs
Fine-tuning FAQs
Multiple API Keys
Dedicated endpoints
Suggest
Deploy a fine-tuned or uploaded LoRA model on serverless for inference