Can I run multiple LLMs on a single dedicated endpoint?

Last updated: November 7, 2025

No, you cannot run multiple models from a single endpoint. Each endpoint is designed to serve one model at a time.