Skip to content
| Marketplace
Sign in
Visual Studio Code>AI>Extra Chat ProvidersNew to Visual Studio Code? Get it now.
Extra Chat Providers

Extra Chat Providers

Cristian Militaru

|
2 installs
| (0) | Free
Extra models providers for Github Copilot Chat
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

🤖 Extra Chat Providers

Bring more AI power to VS Code with additional chat providers

VS Marketplace Version VS Marketplace Downloads
GitHub GitHub Stars License Donate


Extra Chat Providers Icon

✨ Features

  • Xiaomi MiMo: Integration for Xiaomi's AI models directly in your chat. Automatically supports both Pay-as-you-go (sk-...) and Token Plan (tp-...) API keys and routes to the correct endpoints.
  • Z.ai (GLM): Support for Z.ai GLM-4 and GLM-5 models.
  • NVIDIA NIM: OpenAI-compatible support for models from the NVIDIA API Catalog.
  • Native Integration: Works seamlessly with the VS Code LanguageModelChat API.
  • Advanced Capabilities: Supports streaming responses, tool calling, and thinking block rendering.
  • Secure Key Storage: No leaked keys! We use VS Code's built-in Secret Storage for your API keys.
  • Connectivity Testing: Built-in commands to verify your setup works correctly.

📖 Available Models

Xiaomi

Model Context Window Max Output Image Input Tool Calling
MiMo-V2-Pro 1,048,576 131,072 No Yes
MiMo-V2-Flash 262,144 131,072 No Yes
MiMo-V2-Omni 262,144 131,072 Yes Yes

Z.AI

Model Context Window Max Output Image Input Tool Calling
GLM-5.1 204,800 131,072 No Yes
GLM-5V-Turbo 204,800 131,072 Yes Yes
GLM-5 Turbo 204,800 131,072 No Yes
GLM-5 204,800 131,072 No Yes
GLM-4.7 204,800 131,072 No Yes
GLM-4.7 Flash 204,800 131,072 No Yes
GLM-4.6 204,800 131,072 No Yes
GLM-4.5 131,072 98,304 No Yes
GLM-4.5 Air 131,072 98,304 No Yes

NVIDIA NIM

Model Context Window Max Output Image Input Tool Calling
Gemma 4 31B IT (google/gemma-4-31b-it) 262,144 8,192 Yes Yes
MiniMax M2.5 (minimaxai/minimax-m2.5) 204,800 16,384 No Yes
Step 3.5 Flash (stepfun-ai/step-3.5-flash) 262,144 16,384 No Yes
GLM-4.7 (z-ai/glm-4.7) 204,800 16,384 No Yes
Devstral 2 123B Instruct (mistralai/devstral-2-123b-instruct-2512) 262,144 16,384 No Yes
Kimi K2 Instruct 0905 (moonshotai/kimi-k2-instruct-0905) 262,144 16,384 No Yes
Qwen3 Coder 480B (qwen/qwen3-coder-480b-a35b-instruct) 262,144 16,384 No Yes
Kimi K2 Instruct (moonshotai/kimi-k2-instruct) 131,072 16,384 No Yes
Magistral Small 2506 (mistralai/magistral-small-2506) 131,072 16,384 No No
Granite 3.3 8B Instruct (ibm/granite-3.3-8b-instruct) 131,072 8,192 No Yes
QwQ 32B (qwen/qwq-32b) 131,072 32,768 No No
Falcon 3 7B Instruct (tiiuae/falcon3-7b-instruct) 32,768 8,192 No No

🚀 Usage

  1. Install the extension from the VS Code Marketplace or Open VSX.
  2. Open your Language Models panel in VS Code.
  3. Click Add Provider and select Xiaomi, Z.ai, or NVIDIA NIM.
  4. Enter your API key when prompted.
  5. You can now select models from these providers in GitHub Copilot Chat.

Note: VS Code currently exposes extension-contributed language model providers to users on individual GitHub Copilot plans.

⚙️ Commands

  • extra-chat-providers.xiaomi.manage: Manage Xiaomi provider (set/clear API key, test connection).
  • extra-chat-providers.glm.manage: Manage Z.ai (GLM) provider (set/clear API key, test connection).
  • extra-chat-providers.nvidia.manage: Manage NVIDIA NIM provider (set/clear API key, test connection).

🔒 Security & Privacy

  • Local Processing: The extension acts as a bridge between VS Code and the provider APIs.
  • No Mid-man: Your requests go directly to the provider endpoints.
  • Encrypted Keys: API keys are stored in the OS-level keychain via VS Code.

💖 Support the Development

If you find this extension useful, consider buying me a coffee!

Donate with PayPal


🙏 Thank you for using Extra Chat Providers!

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft