Cloud AI
Getting Started
Automation
Reference
Cloud AI (OpenRouter)
Optional cloud backend alongside the built-in local llama engine. Connect to 300+ models with a single API key — disabled by default.
Why Optional Cloud
NeuroTerm ships with a local llama engine that runs entirely on your machine. For most commands and diagnostics, local is fast, private, and free. When you want a bigger model for a harder problem — deep multi-step debugging, complex reasoning, long-context analysis — OpenRouter gives you access to 300+ cloud models through one provider.
The cloud backend is disabled by default. You explicitly opt in per install, and can disable it again at any time.
Setup
Get an OpenRouter API key
Sign up at openrouter.ai and create an API key.
Open Settings → AI → Cloud Backend
Paste the key. It is stored in your OS keychain — Keychain on macOS, Credential Manager on Windows, libsecret on Linux — never written to disk in plain text.
Pick a default model
Choose from frontier models like GPT-4o, Claude Sonnet, Gemini Pro, Qwen, and hundreds more. You can switch models per request.
Use it anywhere AI is available
Neuro Input, Neuro Agent, Neuro Tools, and Auto-Explain all respect the backend choice. A visible indicator shows when cloud AI is active.
Privacy Model
Enabling cloud AI does not send everything to the cloud. Here is exactly what leaves your machine and what stays local:
- API key never leaves your OS keychain
- RAG embeddings stay local even with cloud inference
- Session logs and terminal buffers stay local
- Pattern markers and filters run locally
- Prompts and the context you attach (selected lines, RAG excerpts) go to your chosen provider via OpenRouter
- A visible indicator shows when cloud AI is active
When to Use Cloud vs Local
| Use case | Recommended backend |
|---|---|
| Offline or air-gapped environments | Local only |
| Quick command lookup | Local (faster, free) |
| Privacy-sensitive codebase | Local |
| Deep multi-step debugging | Cloud (bigger models) |
| Long-context datasheet analysis | Cloud (larger context window) |
| Routine boot log review | Local |
Supported Providers
OpenRouter aggregates 300+ models from every major provider. Popular choices include:
- OpenAI — GPT-4o, GPT-4o-mini, o-series reasoning models
- Anthropic — Claude Sonnet, Claude Haiku, Claude Opus
- Google — Gemini Pro, Gemini Flash
- Meta — Llama 3 and Llama 4
- Alibaba — Qwen 2.5 and Qwen 3
- Plus Mistral, DeepSeek, Cohere, and many more
See the full list at openrouter.ai/models.