Use your local Lemonade LLM server with VS Code GitHub Copilot Chat! Connect to your locally running Lemonade server to use local LLM models for code assistance and chat.
🚀 Getting Started
Make sure your Lemonade server is running on http://127.0.0.1:8000
We recommend setting the context size to at least 32k by using the Lemonade tray or --ctx-size cli option.
Install the Lemonade Copilot Chat extension
Open VS Code's chat interface
Click the model picker and click "Manage Models..."
Select "Lemonade" provider
Select from any models that are loaded in your Lemonade server (automatically detected)
Start chatting with your local LLM models! 🥳
NOTE: If needed, configure a custom server URL using the "Manage Lemonade Provider" command
🌟 Why Choose Lemonade for Copilot?
🔒 Complete Privacy: Your code never leaves your machine. Everything stays local and secure
💰 Zero API Costs: No usage fees, no tokens to buy - just pure local AI power
⚡ Lightning Fast: Direct connection to your local server means instant responses
🌐 Works Offline: No internet? No problem! Code assistance anytime, anywhere
🛠️ Advanced Tool Support: Full function calling capabilities for complex tasks
Simple setup: No API keys required
Dynamic model discovery: Automatically detects available models from your server
Requirements
VS Code 1.104.0 or higher
Lemonade server 8.1.10 or higher
🛠️ Development
git clone https://github.com/lemonade-sdk/lemonade-vscode
cd lemonade-vscode
npm install
npm run compile
Press F5 to launch an Extension Development Host.
Common scripts:
Build: npm run compile
Watch: npm run watch
Lint: npm run lint
Format: npm run format
To package the extension, run npx @vscode/vsce package. To install the extension, run code --install-extension lemonade-sdk-0.0.1.vsix.
🔧 Configuration
The extension connects to http://127.0.0.1:8000/api/v1 by default. You can change this by: