If you want to run inference:
Choose from the available models list.
For Serverless Endpoints models, direct inference is possible without the need to initiate a virtual machine (VM).
If you're trying to run inference on a model you fine-tuned, initiate its VM instance either:
Directly from the model's page on api.together.ai, or
Utilizing the start and stop instances of our APIs.
If your desired model isn't listed, feel free to request a model.