Skip to content
| Marketplace
Sign in
Visual Studio Code>Programming Languages>Sentinel CoderNew to Visual Studio Code? Get it now.
Sentinel Coder

Sentinel Coder

Qubitpage

|
3 installs
| (0) | Free
Multi-provider AI coding assistant — 12 providers, 50+ models, local + cloud. Free alternative to GitHub Copilot.
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

Sentinel Coder

Multi-provider AI coding assistant for VS Code — 12 providers, 50+ models, local + cloud. A free, open-source alternative to GitHub Copilot.

Built by QubitPage Research.


Features

  • 12 AI Providers — Ollama (local), Groq (free), Google Gemini (free tier), HuggingFace (free), OpenAI, Anthropic, DeepSeek, Mistral, Together AI, OpenRouter, Vultr, and custom OpenAI-compatible endpoints
  • 50+ Models — GPT-4o, Claude Sonnet 4, Gemini 2.5 Pro, Llama 3.3 70B, Qwen3, DeepSeek V3, and many more
  • Auto Model Router — Automatically picks the best model for each task (coding, reasoning, speed, agentic)
  • Agent Mode — 18 built-in tools: create files, read code, run terminals, search workspace, serve HTML, open browser
  • Thinking Display — See the model's reasoning process in a collapsible panel (for thinking models like o3, DeepSeek R1, Qwen3)
  • Sub-Agent Delegation — Delegate sub-tasks to specialized models (e.g. use a fast model for search, a reasoning model for analysis)
  • Three Modes — Agent (autonomous tools), Ask (Q&A), Plan (analysis without execution)
  • Rich Model Catalog — See pricing, context windows, and feature badges (Tools, Thinking, Vision, Streaming) for every model
  • Zero Telemetry — All API keys stored locally via encrypted SecretStorage. No data leaves your machine except to the providers you choose.

Quick Start

  1. Install from the VS Code Marketplace
  2. Open the Sentinel Coder sidebar (click the S icon in the activity bar)
  3. Pick a provider — Click the ⚙ gear icon → Providers tab → Enable a provider and enter your API key
  4. Select a model from the dropdown
  5. Start chatting!

Free Providers (no credit card needed)

Provider How to get a key Models
Ollama Install Ollama — runs locally, no key needed Any model you pull
Groq console.groq.com — sign up, get free API key Llama 3.3 70B, Qwen3 32B, GPT OSS 120B, Kimi K2, and more
Google Gemini aistudio.google.com — free API key Gemini 2.5 Pro (1M context), Gemini 2.5 Flash
HuggingFace huggingface.co/settings/tokens — free token Qwen 2.5 Coder 32B, Llama 3.3 70B, Mistral Small

Paid Providers

Provider Pricing Highlights
OpenAI Pay-per-use GPT-4o, o3, o4-mini, Codex Mini
Anthropic Pay-per-use Claude Opus 4, Claude Sonnet 4
DeepSeek Pay-per-use (very cheap) DeepSeek V3, DeepSeek R1
Mistral Free tier + paid Mistral Large, Codestral
Together AI Pay-per-use Llama 3.1 405B, Qwen 2.5 Coder 32B
OpenRouter Pay-per-use (aggregator) Routes to cheapest provider
Vultr Pay-per-use DeepSeek V3.2, GLM 5.1, Kimi K2.5

Modes

Agent Mode

The AI can autonomously use tools to complete tasks:

  • File Tools — createFile, readFile, listDirectory, searchFiles, searchText
  • Editor Tools — getActiveFile, getSelection, getDiagnostics, openFile
  • Terminal Tools — runCommand
  • Web Tools — serveFile, openBrowser, fetchUrl
  • Analysis Tools — delegateSubAgent (spawn sub-tasks on different models)

Ask Mode

Direct Q&A — the AI answers questions and writes code without using tools.

Plan Mode

The AI analyzes your request and creates a step-by-step plan without executing anything.

Configuration

All settings are under sentinelCoder.* in VS Code settings:

Setting Default Description
sentinelCoder.ollamaUrl http://127.0.0.1:11434 Ollama API URL
sentinelCoder.model sentinel-coder:latest Default Ollama model
sentinelCoder.maxTokens 2048 Max tokens to generate
sentinelCoder.temperature 0.3 Generation temperature (0-1)
sentinelCoder.defaultMode agent Default mode (agent/ask/plan)
sentinelCoder.approvalMode default Tool approval (default/bypass/autopilot)
sentinelCoder.apiKeysFile (empty) Optional: path to bulk-import API keys from a text file

API Keys File Format (Optional)

For power users who want to bulk-import keys from a text file:

groq: gsk_your_groq_key_here
openrouter: sk-or-v1-your_openrouter_key
gemini: AIzaSyYour_google_key
vultr inference: YOUR_VULTR_KEY
huggingface: hf_your_hf_token

Set sentinelCoder.apiKeysFile to the file path. Keys entered via the Settings > Providers panel are stored securely in VS Code's encrypted SecretStorage each.

Privacy & Security

  • Zero telemetry — No analytics, no tracking, no data collection
  • API keys encrypted — Stored in VS Code SecretStorage (OS-level encryption), never in plaintext settings
  • Local-first — Works fully offline with Ollama. Cloud providers are optional
  • Open source — MIT License. Inspect every line of code

Requirements

  • VS Code 1.93.0 or later
  • For local models: Ollama installed and running
  • For cloud models: An API key from any supported provider

Known Issues

  • The Chat Participant (@sentinel in the native chat panel) requires VS Code 1.93+
  • Some thinking models may have a delay before the first response token

Release Notes

3.0.0

  • Multi-provider support (12 providers, 50+ models)
  • Auto model router — picks the best model per task
  • Sub-agent delegation for complex workflows
  • Thinking display for reasoning models
  • Rich model catalog with pricing and feature badges
  • Secure API key storage via SecretStorage
  • CoderQ personality mode

Made with ❤️ by QubitPage Research

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft