Docs/Cloud AI

Cloud AI (OpenRouter)

Optional cloud backend alongside the built-in local llama engine. Connect to 300+ models with a single API key — disabled by default.

Why Optional Cloud

NeuroTerm ships with a local llama engine that runs entirely on your machine. For most commands and diagnostics, local is fast, private, and free. When you want a bigger model for a harder problem — deep multi-step debugging, complex reasoning, long-context analysis — OpenRouter gives you access to 300+ cloud models through one provider.

The cloud backend is disabled by default. You explicitly opt in per install, and can disable it again at any time.

Setup

1

Get an OpenRouter API key

Sign up at openrouter.ai and create an API key.

2

Open Settings → AI → Cloud Backend

Paste the key. It is stored in your OS keychain — Keychain on macOS, Credential Manager on Windows, libsecret on Linux — never written to disk in plain text.

3

Pick a default model

Choose from frontier models like GPT-4o, Claude Sonnet, Gemini Pro, Qwen, and hundreds more. You can switch models per request.

4

Use it anywhere AI is available

Neuro Input, Neuro Agent, Neuro Tools, and Auto-Explain all respect the backend choice. A visible indicator shows when cloud AI is active.

Privacy Model

Enabling cloud AI does not send everything to the cloud. Here is exactly what leaves your machine and what stays local:

  • API key never leaves your OS keychain
  • RAG embeddings stay local even with cloud inference
  • Session logs and terminal buffers stay local
  • Pattern markers and filters run locally
  • Prompts and the context you attach (selected lines, RAG excerpts) go to your chosen provider via OpenRouter
  • A visible indicator shows when cloud AI is active

When to Use Cloud vs Local

Use caseRecommended backend
Offline or air-gapped environmentsLocal only
Quick command lookupLocal (faster, free)
Privacy-sensitive codebaseLocal
Deep multi-step debuggingCloud (bigger models)
Long-context datasheet analysisCloud (larger context window)
Routine boot log reviewLocal

Supported Providers

OpenRouter aggregates 300+ models from every major provider. Popular choices include:

  • OpenAI — GPT-4o, GPT-4o-mini, o-series reasoning models
  • Anthropic — Claude Sonnet, Claude Haiku, Claude Opus
  • Google — Gemini Pro, Gemini Flash
  • Meta — Llama 3 and Llama 4
  • Alibaba — Qwen 2.5 and Qwen 3
  • Plus Mistral, DeepSeek, Cohere, and many more

See the full list at openrouter.ai/models.

Related