Prerequisites
- Together API Key – Required for all operations. Get one from together.ai.
- Dedicated Containers access – Contact your account representative or [email protected] to enable Dedicated Containers for your organization.
- Docker – For building and pushing container images. Get it here.
- uv (optional) – For Python/package management. Install from astral-sh/uv.
Step 1: Install the Together CLI
Step 2: Clone the Sprocket Examples
Step 3: Build and Deploy
Navigate to the example worker and deploy:- Builds the Docker image from the example
- Pushes it to Together’s private registry
- Creates a deployment on Together’s GPU infrastructure
pyproject.toml and from the output (you’ll need it for the next steps).
The example worker uses this pyproject.toml configuration:
Step 4: Watch Deployment Status
running and replicas are ready. Press Ctrl+C to stop watching. Note that watch is not installed by default on MacOS, use brew install watch or your package manager of choice.
Step 5: Test the Health Endpoint
Replace<your-deployment> with your deployment name:
Step 6: Submit a Job
Replace<your-deployment> with your deployment name:
request_id for the next step.
Step 7: Get the Job Result
Replace<your-deployment> and <request-id> with your values:
multiplicand: 5 returns value: 10.
Step 8: View Logs
Stream logs from your deployment:Step 9: Clean Up
When you’re done, delete the deployment:Next Steps
Now that you’ve deployed your first container, explore the full platform:- Dedicated Containers Overview – Architecture and concepts
- Jig CLI – Build, push, deploy, secrets, and volumes
- Sprocket SDK – Build queue-integrated inference workers
- API Reference – REST API for deployments, secrets, and queues
Example Guides
- Image Generation with Flux2 – Single-GPU inference with 4-bit quantization
- Video Generation with Wan 2.1 – Multi-GPU inference with torchrun