Skip to main content
All CollectionsInference FAQs
I encountered an error while using your API. What should I do?
I encountered an error while using your API. What should I do?
Updated over a week ago

Code

Cause

Solution

401 - Authentication Error

Missing or Invalid API Key

Ensure you are using the correct API Key and supplying it correctly

403 - Bad Request

Input token count + max_tokens parameter must be less than the context length of the model being queried.

Set max_tokens to a lower number. If querying a chat model, you may set max_tokens to null and let the model decide when to stop generation.

404 - Not Found

Invalid Endpoint URL or model name

Check your request is being made to the correct endpoint (see the API reference page for details) and that the model being queried is available

429 - Rate limit

Too many requests sent in a short period of time

Throttle the rate at which requests are sent to our servers (see our rate limits)

500 - Invalid Request

Misconfigured request

Ensure your request is a Valid JSON and your API Key is correct. Also ensure you're using the right prompt format - which is different for Mistral and LLaMA models.

503 - Engine Overloaded

Our servers are seeing high amounts of traffic

Please try again after a brief wait. If the the issue persists, please contact support

504 - Timeout

Our servers are overloaded with requests and could not process your request on time

Please try again after a brief wait. If the the issue persists, please contact support

524 - Timeout

Cloudflare established a connection to the server it's supposed to communicate with, but the server took too long to respond.

Please try again after a brief wait. If the the issue persists, please contact support

529 - Server Error

Error on our servers

This is an issue from our end. Please try again after a wait. If the issue persists, please contact support

If you are seeing other error codes or the solutions do not work, please contact support for help.

Did this answer your question?