Llama.cpp Provider for GitHub Copilot ChatThis extension integrates Llama.cpp models into GitHub Copilot Chat in VS Code. Features
Requirements
Stack
DesignThe extension uses a base provider class for OpenAI-compatible chat APIs. The Llama.cpp provider extends this base to connect to a local Llama.cpp server. It handles model fetching, message conversion, and streaming responses. Tool calling is supported through OpenAI-compatible formats. Setting Up Project
UsageInstall the extension from the marketplace. Configure the Llama.cpp server URL via the command palette. Select the Llama.cpp provider in the chat interface. Start chatting with the integrated models. References |