Skip to content
| Marketplace
Sign in
Visual Studio Code>Machine Learning>LiteLLM Connector for CopilotNew to Visual Studio Code? Get it now.
LiteLLM Connector for Copilot

LiteLLM Connector for Copilot

Gethnet

|
546 installs
| (0) | Free
| Sponsor
An extension that integrates LiteLLM proxy into Copilot
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

🚀 LiteLLM Connector for GitHub Copilot Chat

CI Codecov GitHub release (latest SemVer) Open VSX Version Open VSX Downloads

Bring any LiteLLM-supported model into the Copilot Chat model picker — OpenAI, Anthropic (Claude), Google, Mistral, local Llama, and more.

If LiteLLM can talk to it, Copilot can use it.


⭐️ Support the project

If you find this useful, please:

  • Star on GitHub: https://github.com/gethnet/litellm-connector-copilot
  • Leave a rating/review on the VS Code Marketplace: https://marketplace.visualstudio.com/items?itemName=GethNet.litellm-connector-copilot
  • Rate on Open VSX: https://open-vsx.org/extension/GethNet/litellm-connector-copilot

Want to support development?

  • Ko-fi: https://ko-fi.com/amwdrizz
  • Buy Me a Coffee: https://buymeacoffee.com/amwdrizz

🚨 Troubleshooting: Connection Issues 🚨

If the extension fails to connect or models don't show up:

  1. Manual Setup: Run Manage LiteLLM Provider from the Command Palette (Ctrl+Shift+P). This often fixes setup "hiccups".
  2. Verify: Run LiteLLM: Check Connection to test your settings.
  3. Reset: If things are totally stuck, run LiteLLM: Reset All Configuration. This is the "nuke" option to clear all state.
  4. Note: Reinstalling usually won't help as settings are stored securely in VS Code. Use the Reset command instead.

✅ Requirements

  • �️ VS Code 1.110+ (required)
  • �🔑 GitHub Copilot subscription (Free plan works)
  • 🌐 A LiteLLM proxy URL (and an API key if your proxy requires one)

⚡ Quick Start (60 seconds)

  1. Install GitHub Copilot Chat
  2. Install LiteLLM Connector for Copilot
  3. Open Command Palette: Ctrl+Shift+P / Cmd+Shift+P
  4. Run: Manage LiteLLM Provider
  5. Enter:
    • Base URL (example: http://localhost:4000)
    • API Key (optional)
  6. Open Copilot Chat → pick a model under LiteLLM → chat

Multi-Backend: Connect to multiple LiteLLM instances simultaneously. Use Manage Multiple Backends to aggregate models from several proxies. Models are automatically namespaced (e.g., local/llama-3).


✨ What you get

  • 🌍 Hundreds of models via your LiteLLM proxy
  • ⛓️ Multi-backend aggregation from multiple proxy instances
  • 🌊 Real-time streaming responses
  • 🛠️ Tool / function calling support
  • 👁️ Vision models supported (where available)
  • 🧠 V2 Chat Provider (Experimental) — supports thinking parts for reasoning models via newer VS Code APIs
  • 🧠 Smart parameter handling for model quirks
  • 🔁 Automatic retry when a model rejects unsupported flags
  • 📊 Token tracking & usage monitoring for input/output tokens
  • ✍️ Git commit generation from staged changes in the SCM view (works in multi-repo workspaces)
  • 🧼 Smart Sanitization automatically strips Markdown code blocks from generated commit messages
  • 🔍 Connection diagnostics to verify proxy configuration
  • ⏱️ Inactivity watchdog to prevent stuck streams
  • 🚫🧠 Cache bypass controls (no-cache headers) with provider-aware behavior
  • 🔐 Secure credential storage using VS Code SecretStorage
  • ⌨️ Optional inline completions via VS Code's stable inline completion API

🆕 Recent Highlights

  • 📝 Multi-Repo Commit Generation (correctly identifies the active repository in multi-repo workspaces)
  • 🧪 Telemetry & Observability (PostHog-backed feature tracking and structured JSONL logging)
  • 🔧 Model Capability Overrides (manually override toolCalling and imageInput detection)
  • 🧠 V2 Chat Provider (experimental support for thinking parts and newer VS Code chat APIs)
  • 🧼 SCM Message Sanitization (automatically cleans up generated commit messages by stripping triple backticks)
  • ✍️ Git Commit Message Generation (generate messages from staged changes directly in the SCM view)
  • 📊 Enhanced Token Awareness (real-time token counting and context window display in model tooltips)
  • 🔍 Connection Diagnostics (new Check Connection command to validate proxy settings)
  • 🚀 VS Code 1.110+ settings modernization (aligns with the Language Model provider settings UI)
  • 🧱 Tool-call compatibility hardening (normalizes tool call IDs to OpenAI-compatible limits)
  • 🧰 Stability Improvements (hardened JSON parsing and stream error recovery)
  • 📦 Smaller, faster package (bundled/minified production builds)

⚙️ Configuration

Setting Type Default Description
litellm-connector.baseUrl string "" Base URL for the LiteLLM proxy server.
litellm-connector.backends array [] List of LiteLLM backends to connect to.
litellm-connector.inactivityTimeout number 60 Seconds of inactivity before the connection is considered idle.
litellm-connector.disableCaching boolean true Send no-cache headers to bypass LiteLLM caching.
litellm-connector.commitModelIdOverride string "" Override the model used for git commit message generation.
litellm-connector.enableResponsesApi boolean false (Experimental) Enable the VSCode Responses API integration.
litellm-connector.disableQuotaToolRedaction boolean false Disable automatic tool removal when a quota error is detected.
litellm-connector.modelOverrides object {} Override or add tags for specific models.
litellm-connector.modelCapabilitiesOverrides object {} Override model capabilities (toolCalling, imageInput) reported to VS Code (e.g., toolCalling,imageInput).
litellm-connector.inlineCompletions.enabled boolean false Enable LiteLLM inline completions. (Deprecated: will be removed)
litellm-connector.inlineCompletions.modelId string "" (Deprecated) Use VS Code's inlineChat.defaultModel setting instead.
litellm-connector.emitUsageData boolean false (Experimental) Emit token usage metadata.
litellm-connector.sendDefaultParameters boolean false (Temporary, will be removed) Send default parameters if not provided. Recommended: false.

⌨️ Commands

  • Manage LiteLLM Provider: Configure Base URL + API Key; refreshes models.
  • LiteLLM: Check Connection: Verify proxy URL and API key configuration.
  • LiteLLM: Select Inline Completion Model: Choose a model for inline completions.
  • LiteLLM: Select Commit Message Model: Choose a model for git commit generation.
  • LiteLLM: Show Available Models: See all models currently discovered from your proxy.
  • LiteLLM: Reload Models: Manually refresh the model list.
  • LiteLLM: Reset All Configuration: Clear all stored URLs and API keys.

🐛 Bug reports & feature requests

Please use GitHub Issues: https://github.com/gethnet/litellm-connector-copilot/issues

Including VS Code version, extension version, model id, and LiteLLM proxy details/logs (if possible) helps reproduce issues quickly.


🧩 Notes

  • This extension is a provider for the official Copilot Chat experience.
  • It won't function without the GitHub Copilot Chat extension installed.

🆘 Support

  • Issues & feedback: https://github.com/gethnet/litellm-connector-copilot/issues
  • License: Apache-2.0
  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft