Skip to content
| Marketplace
Sign in
Visual Studio Code>Programming Languages>MiMo for Copilot ChatNew to Visual Studio Code? Get it now.
MiMo for Copilot Chat

MiMo for Copilot Chat

clockzinc_bit

|
8 installs
| (0) | Free
Use MiMo V2.5 Pro & DeepSeek V4 models in Copilot Chat. Vision, thinking mode, agent tools. Multi-provider BYOK.
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

MiMo for Copilot Chat

Forked from Vizards/deepseek-v4-for-copilot — original extension by Vizards. This fork adds Xiaomi MiMo V2.5 Pro model support, multi-provider configuration UI, and enhanced model management.

GitHub

Use MiMo V2.5 Pro and DeepSeek V4 models from the Copilot Chat model picker — and keep everything else Copilot already gives you.

DeepSeek V4 Pro and Flash in the Copilot Chat model picker, with the per-model Thinking Effort dropdown (None / High / Max)

Love DeepSeek's price-performance but don't want to give up GitHub Copilot's agent mode, tool calling, and polished UI? This extension drops DeepSeek V4 Pro & Flash straight into the Copilot Chat model selector — with vision, thinking mode, and your own API key.

Why this extension?

  • Don't replace Copilot — power it up. No new sidebar, no new chat UI to learn. Just a new model in the picker you already use.
  • Agent mode, tool calling, instructions, MCP, skills — all of it still works. Copilot's entire stack, now running on DeepSeek.
  • Vision on a text-only model. DeepSeek V4 can't see images. This extension proxies any image you drop into chat through another Copilot model you already have, then feeds the description to DeepSeek — transparently.
  • BYOK, pay DeepSeek directly. Your API key, your bill, your rate limits. Stored in the OS keychain, never on disk.

Features

DeepSeek V4 Pro & Flash in the model picker

Both models show up alongside GPT-4o, Claude, and friends in Copilot Chat's model selector. 1M token context on both. Switch models mid-chat without losing history.

Transparent Vision Proxy

DeepSeek V4 is text-only. Drop a screenshot into chat and this extension automatically hands the image to another installed Copilot model (Claude, GPT-4o, whatever you've got), gets a description, and feeds that back to DeepSeek. Zero config — just pick your preferred vision model once.

Dropping an image into Copilot Chat and DeepSeek responding to it via the vision proxy

Thinking Mode with Reasoning Effort Control

Full support for DeepSeek V4's reasoning_content. Use Copilot Chat's native model picker menu to choose none (off), high (balanced, default), or max (deep reasoning for hard agent tasks).

Inherits Every Copilot Capability

Because this plugs into Copilot's native provider API, you get the full stack for free:

  • Agent mode — autonomous multi-step tasks
  • Tool calling — file edits, terminal, workspace search, Git, tests
  • Instructions & skills — all your .instructions.md, AGENTS.md, and skills just work
  • Prompt caching stats — DeepSeek's cache hit rate logged in the output channel so you can see the savings

DeepSeek V4 Pro running Copilot's agent mode with tool calls

Secure by Default

API key lives in VS Code's SecretStorage (OS keychain on macOS / Windows / Linux). Never in settings.json, never in your Git history.

Zero Runtime Dependencies

Pure VS Code API + Node.js built-ins. No Python, no Docker, no local proxy server to babysit.

Getting Started

Prerequisites

  • VS Code 1.116 or later. This extension relies on non-public Copilot Chat APIs that may break on newer VS Code versions — report an issue if you hit one.
  • GitHub Copilot subscription (Free / Pro / Enterprise — the free tier works)
  • DeepSeek API key from platform.deepseek.com

Usage

  1. Install from the VS Code Marketplace
  2. Run DeepSeek: Set API Key from the Command Palette (Cmd+Shift+P)
  3. Paste your key (starts with sk-)
  4. Open Copilot Chat, click the model picker, pick DeepSeek V4 Pro or DeepSeek V4 Flash
  5. That's it — chat away

Models

Model Best For
DeepSeek V4 Flash Fast everyday coding, quick edits, cheap iteration
DeepSeek V4 Pro Complex refactors, agent tasks, deep reasoning

Both support optional thinking mode, tool calling, and 1M token context.

Settings

Setting Default Description
deepseek-copilot.baseUrl https://api.deepseek.com API endpoint — change for self-hosted / proxied deployments
deepseek-copilot.maxTokens 0 Max output tokens (0 = no limit). Useful for cost control
deepseek-copilot.modelIdOverrides prefilled official ID map API model IDs to send for DeepSeek V4 Flash / Pro. Change only for compatible third-party APIs with different model names
deepseek-copilot.visionModel (auto) Which Copilot model to proxy images through
deepseek-copilot.visionPrompt (built-in) Prompt used to describe image attachments

Thinking Effort is configured from Copilot Chat's model picker for each DeepSeek model.

Example settings.json override for compatible API proxies:

{
  "deepseek-copilot.modelIdOverrides": {
    "deepseek-v4-flash": "your-flash-model-id",
    "deepseek-v4-pro": "your-pro-model-id"
  }
}

Compared to alternatives

This extension Local proxy (e.g. LiteLLM) Standalone DeepSeek extensions
Works inside Copilot Chat ✅ ✅ ❌ separate UI
Agent mode, tools, skills ✅ ✅ ⚠️ reimplemented
Vision support ✅ proxied ❌ ❌
Extra process to run ❌ ✅ ❌
One-click install ✅ ❌ ✅
API key in OS keychain ✅ ❌ ⚠️ varies

License

MIT

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft