Skip to main content
Dedicated Containers let you run your own Dockerized inference workloads on Together’s managed GPU infrastructure. You bring the container — Together handles compute provisioning, autoscaling, networking, and observability. You build and push a Docker image using the Jig CLI. Inside your container, the Sprocket SDK connects your inference code to Together’s managed job queue. Once deployed, your workers can receive requests.
  • Wrap and deploy your model in 20 minutes
  • Boost conversion and margins with fair priority queueing
  • Bottomless capacity just before you need it
Dedicated Containers Architecture

Quickstart

Deploy Your First Container

Deploy your first container from the command line

Concepts

Platform Overview

Architecture, deployment lifecycle, autoscaling, and troubleshooting

Guides

Reference


Get Access

Contact your account representative or [email protected] to enable Dedicated Containers for your organization.