The Vast.ai CLI automatically retries rate-limited requests. You do not need to implement your own retry logic — the CLI handles HTTP 429 responses with exponential backoff out of the box. This page covers the error format, how rate limits work, and how to configure the CLI’s built-in retry behavior. For the full details on rate limit mechanics, see the API Rate Limits and Errors page.Documentation Index
Fetch the complete documentation index at: https://docs.vast.ai/llms.txt
Use this file to discover all available pages before exploring further.
Error Responses
When a CLI command fails, it prints the error message from the API and exits with a non-zero status code. The underlying API error shape is:success or error and return only msg or message. The CLI surfaces whatever message the API returns.
How Rate Limits Work
Vast.ai applies rate limits per endpoint and per identity. The identity is determined by your bearer token, session user,api_key parameter, and client IP.
Some endpoints also enforce method-specific limits (GET vs POST) and max-calls-per-period limits for short bursts.
For the full breakdown, see How rate limits are applied.
Rate Limit Response
When you hit a rate limit, the API returns HTTP 429 with a message like:Retry-After header. The CLI handles this automatically using its built-in retry logic.
Built-in Retry Behavior
When the CLI receives an HTTP 429 response, it automatically retries the request using exponential backoff:- Retried status codes: 429 only
- Default retries: 3
- Backoff strategy: starts at 0.15 seconds, multiplied by 1.5x after each attempt
- Retry delays: ~0.15s, ~0.225s, ~0.34s
Configuring Retries
Use the--retry flag on any command to change the number of retry attempts: