Free & open-source universal API Gateway for all VS Code language models. Auto-discovers GitHub Copilot, Gemini, Ollama & any AI extension — exposing them via local OpenAI, Anthropic, Google & Llama compatible APIs. One endpoint, every model. Use with Cursor, LangChain, Agents & more.
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Use GitHub Copilot, Gemini, Ollama & any VS Code AI — like any other API.
One extension. Zero API keys. Auto-discovers every model. Works with LangChain, Cursor, Aider, and 50+ tools.
⚡ Quick Start
# 1. Install from VS Code Marketplace (search "GitHub Copilot API Gateway")
# 2. Click "Start Server" in the sidebar
# 3. Done. Your local API is ready.
Use it like OpenAI:
from openai import OpenAI
client = OpenAI(
base_url="http://127.0.0.1:3030/v1",
api_key="anything" # No real API key needed
)
response = client.chat.completions.create(
model="gpt-4o",
messages=[{"role": "user", "content": "Hello!"}]
)
🔒 Your data stays local. This extension runs entirely on your machine — no third-party servers, no telemetry, no data collection. The full source code is available on GitHub for you to audit.
✅ What You Can Do
Use Cursor with Copilot, Gemini, or Ollama as the backend
All models are exposed at GET /v1/models and routable via the model field in any request. No config changes needed — just install the VS Code extension and the models appear.
🤖 Using with Clawdbot (Moltbot)
Clawdbot is an open-source AI assistant that brings LLMs to messaging platforms like WhatsApp, Telegram, Discord, and Slack. Since it supports any OpenAI-compatible API endpoint, you can use this gateway to power Clawdbot with GitHub Copilot:
Start the Copilot API Gateway in VS Code
Expose the API (set host to 0.0.0.0 if Clawdbot runs on another machine)
Configure Clawdbot to use your gateway:
# In your Clawdbot config
llm:
provider: openai
base_url: http://YOUR-IP:3030/v1
api_key: your-optional-api-key # or "anything" if auth is disabled
model: gpt-4o
Now your WhatsApp/Telegram/Discord/Slack bots are powered by Copilot! 🚀
🌐 Internet Access via Cloudflare Tunnels
Access your Copilot API from anywhere — your phone, tablet, or share with friends:
Enable authentication (set an API key in Security settings)
Start the server
Click "Go Live" in the Internet Access section
Get a public *.trycloudflare.com URL instantly
Free, no Cloudflare account needed! Perfect for:
Accessing from your phone outside your home network
Sharing with friends or teammates temporarily
Remote development without VPN setup
🖥️ Run as a Background Service
The gateway runs inside VS Code's extension host, so VS Code must be running — but it doesn't need to be visible. Here's how to run it like a background service:
Prerequisites: Enable autoStart so the server launches automatically:
{ "githubCopilotApi.server.autoStart": true }
macOS:
# Launch VS Code minimized (gateway starts automatically)
open -a "Visual Studio Code" --args --disable-workspace-trust
# Or run completely headless via CLI
nohup code --disable-workspace-trust &>/dev/null &
Windows (PowerShell):
# Launch minimized
Start-Process code -ArgumentList "--disable-workspace-trust" -WindowStyle Minimized
# Or add to Task Scheduler for start-on-boot:
# Program: code
# Arguments: --disable-workspace-trust
Linux:
# Run in background
nohup code --disable-workspace-trust &>/dev/null &
# Or create a systemd user service (~/.config/systemd/user/copilot-api.service):
# [Service]
# ExecStart=/usr/bin/code --disable-workspace-trust
# Restart=on-failure
# Then: systemctl --user enable --now copilot-api
Note: VS Code must remain running in the background. If you close VS Code, the gateway stops. The autoStart setting ensures the API server starts the moment VS Code opens — no manual clicks needed.
🔌 API Endpoints
Endpoint
Method
Description
/v1/chat/completions
POST
Chat completions (streaming supported)
/v1/completions
POST
Legacy completions API (streaming supported)
/v1/responses
POST
OpenAI Responses API (2026 spec, streaming supported)
/v1/models
GET
List all available models
/v1/tools
GET
List available tools (VS Code + MCP)
/v1/tools/call
POST
Execute a tool directly
/v1/mcp/servers
GET
List connected MCP servers
/v1/messages
POST
Anthropic Claude-compatible endpoint
/v1beta/models/:model:generateContent
POST
Google Gemini-compatible endpoint
/health
GET
Server health check
/docs
GET
Interactive Swagger UI
⚙️ Configuration
Customize in VS Code Settings (githubCopilotApi.*):
🚫 No telemetry — zero data collection, no analytics, no tracking
🔒 Runs locally — your prompts and API keys never leave your machine
✅ Community-driven — contributions welcome via CONTRIBUTING.md
🛡️ Security — report vulnerabilities via SECURITY.md
Disclaimer: This extension is an independent, community-built project and is not affiliated with GitHub, Microsoft, or OpenAI. It leverages your existing GitHub Copilot subscription. Use responsibly.