Redirect cloud AI API traffic (OpenAI, Anthropic, Gemini) to a local Ollama instance or another backend, transparently to the application.
localhost:11434Press R to open the Inference Routing panel.
cella comes with built-in presets:
| Source | Backend | Description |
|---|---|---|
api.openai.com | localhost:11434 | OpenAI → local Ollama |
api.anthropic.com | localhost:11434 | Anthropic → local Ollama |
api.business.githubcopilot.com | localhost:11434 | Copilot → local Ollama |
generativelanguage.googleapis.com | localhost:11434 | Gemini → local Ollama |
Navigate to the desired route and press Enter to toggle it on. The route takes effect immediately.
When the MITM proxy intercepts a request for a routed domain, it rewrites the upstream destination to the configured backend. The application sees a normal HTTPS response — no code changes needed.
/v1) handles requests from OpenAI, Copilot, and Gemini clients natively. The Anthropic route needs an API format adapter since Anthropic uses a different request schema.