Setup
See our Getting Started guide for initial setup.Upload
You can upload models from Hugging Face or S3 and run inference on a dedicated endpoint through Together AI. For more information, see our Dedicated Inference docsOptions
| Name | Arguments | Description |
|---|---|---|
--model-name | string | The name to give to your uploaded model [required] |
--model-source | string | The source uri of the model [required] |
--model-type | model or adapter | Whether the model is a full model or an adapter |
--hf-token | string | Hugging Face token (if uploading from Hugging Face) |
--description | string | A description of your model |
--base-model | string | The base model to use for an adapter if setting it to run against a serverless pool. Only used for model_type ‘adapter’. |
--lora-model | string | The lora pool to use for an adapter if setting it to run against, say, a dedicated pool. Only used for model_type ‘adapter’. |
--json | Output in JSON format |
List all models
Options
| Name | Description |
|---|---|
--type | Filter models by type. dedicated is the only available option |
--json | Output in JSON format |