Skip to content
| Marketplace
Sign in
Visual Studio Code>Other>EasyCode - Local AI ChatNew to Visual Studio Code? Get it now.
EasyCode - Local AI Chat

EasyCode - Local AI Chat

brunoprela

|
1 install
| (0) | Free
AI chat assistant powered by local models (similar to Cursor)
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

EasyCode - Local AI Chat for VSCode

A VSCode extension that brings a Cursor-like chat interface powered by local AI models (Ollama).

Features

  • 🚀 Cursor-like Chat Interface: Clean, modern chat UI similar to Cursor
  • 🤖 Local AI Models: Uses Ollama for local model inference (no cloud dependency)
  • 💻 Code Context Awareness: Automatically includes code context from your open files
  • 🎨 VS Code Theme Integration: Adapts to your VS Code theme
  • ⚡ Fast & Responsive: Lightweight and fast

Prerequisites

  1. Ollama: Install Ollama and make sure it's running
  2. Models: Pull at least one model using Ollama:
    ollama pull llama2
    # or
    ollama pull codellama
    # or
    ollama pull mistral
    

Installation

  1. Clone this repository
  2. Install dependencies:
    npm install
    
  3. Compile the extension:
    npm run compile
    
  4. Press F5 in VS Code to open a new Extension Development Host window
  5. In the new window, press Ctrl+Shift+L (or Cmd+Shift+L on Mac) to open the chat panel

Usage

  1. Open Chat Panel:

    • Press Ctrl+Shift+L (or Cmd+Shift+L on Mac)
    • Or use Command Palette: "Open EasyCode Chat"
  2. Select Model: Choose from available Ollama models in the dropdown

  3. Chat: Type your message and press Enter (Shift+Enter for new line)

  4. Code Context: The extension automatically includes context from:

    • Currently open file
    • Selected code (if any)
    • File language and path

Configuration

You can configure the Ollama URL in VS Code settings:

  • Open Settings (Ctrl+, or Cmd+,)
  • Search for "EasyCode"
  • Set EasyCode: Ollama Url (default: http://localhost:11434)

Development

# Install dependencies
npm install

# Compile TypeScript
npm run compile

# Watch mode (auto-compile on changes)
npm run watch

# Lint
npm run lint

Requirements

  • VS Code 1.74.0 or higher
  • Node.js 18.x or higher
  • npm 10.x or higher
  • Ollama running locally

License

MIT

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2025 Microsoft