Skip to content
| Marketplace
Sign in
Visual Studio Code>Other>ollama-viewNew to Visual Studio Code? Get it now.
ollama-view

ollama-view

Oswaldo Dantas

|
7 installs
| (0) | Free
Manage local Ollama models
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

Ollama View for VS Code

VS Code Marketplace

Manage your locally installed Ollama models directly from Visual Studio Code.

(Experimental) Start quick experiments by chatting directly with your models.

Ollama View Screenshot

Installation

You can install this extension from the following sources:

  • VS Code Marketplace: Official release.
  • ~~Open VSX Registry: For Codium and other editors~~ (coming soon).
  • GitHub Releases: Download the .vsix file manually.

Features

Sidepanel View

  • Model List: View all your local Ollama models in a tree view.
  • Pull Models: Easily download new models (e.g., llama3, mistral) via command.
  • Status Indication: Visual indicators for Running (Green) and Stopped (Grey) models.
  • Persistent Chats: Create and manage multiple persistent chats for each model.
  • Context Menu: Right-click to Start, Stop, or Delete models and chats.

Chat Interface

  • Concurrent Chats: Open multiple chat tabs simultaneously.
  • History: Chat history is automatically saved and persists across VS Code sessions.
  • Auto-Start: Sending a message to a stopped model continuously starts it.
  • Message Editing: Hover over user messages to copy content or access edit options:
    • Edit / Truncate: Edit a message and remove all subsequent history (rewriting the conversation path).
    • Edit / Fork: Edit a message and create a new chat branch, preserving the original conversation.
  • Model Actions: Hover over model responses to:
    • Copy: Copy response content to clipboard.
    • Regenerate: Re-generate the answer. If it's not the last message, this will truncate the conversation.
    • Fork: Create a new chat branching from this point.
  • Context-Aware Commands: Run Start/Stop/Delete from the Command Palette (Ctrl+Shift+P) to see a interactive list of models if you haven't selected one in the view.

Requirements

  • Ollama must be installed and running locally.
  • By default, it connects to http://127.0.0.1:11434.

Note: This version (v0.0.1) has been tested primarily on Linux. It assumes Ollama is already installed and running correctly on your system. Compatibility with other OSs is expected but not yet verified.

Commands

  • ollama-view: Refresh: Refresh the model list.
  • ollama-view: Pull Model: Download a new model.
  • ollama-view: Start: Start a model (context-aware).
  • ollama-view: Stop: Stop a running model (context-aware).
  • ollama-view: Delete: Delete a model (context-aware).

Known Issues

  • Autocomplete for model names during "Pull" is limited to a popular subset. You can still type any model name manually.

Release Notes

0.0.3 - Unique Chat Naming

  • Unique Chat Names: New chats now have unique names (e.g. "New Chat (2)") to prevent confusion.
  • Improved Naming Logic: Renaming and forking chats also ensures unique names.
  • Tests: Added tests for chat naming logic.

0.0.2 - Initial Chat Functionality

  • Persistent Chats: Conversations are now possible and persistent across sessions.
  • Message Editing: Edit user messages to branch conversation paths (Truncate or Fork).
  • Model Actions: Copy, regenerate, and fork directly from model responses.
  • Enhanced UI: Improved tree view, chat deletion flow, and timestamp formatting.

0.0.1 - Foundations

  • Initial release with View, Start, Stop, Delete, and Pull functionalities.
  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft