Base URL
your-jarvis-host with the hostname or IP of your Jarvis node. All endpoints are served over HTTPS.
Authentication
Every request must include an API key in theAuthorization header.
Your Jarvis API key, prefixed with
Bearer. Example: Bearer jrv_yourkey123.OpenAI compatibility
The Jarvis model API is OpenAI-compatible via LiteLLM. Any client that works with the OpenAI API — including the official Python and Node.js SDKs, LangChain, and LlamaIndex — works with Jarvis. Point the client at your Jarvis host and supply your Jarvis API key.
Rate limiting
Jarvis does not enforce hard rate limits by default, but available throughput depends on your hardware and how many models are loaded. Best practices:- Reuse HTTP connections — avoid opening a new connection per request.
- Use the model that fits your task. Smaller models like
ollama/mistralrespond faster for simple tasks. - Monitor GPU utilization via the monitoring guide to spot saturation early.
Error responses
All errors return a JSON body with a consistent structure.The top-level error container.