Skip to content
| Marketplace
Sign in
Visual Studio Code>Azure>Azure OpenAI PromptOps ManagerNew to Visual Studio Code? Get it now.
Azure OpenAI PromptOps Manager

Azure OpenAI PromptOps Manager

Shas Vaddi

|
2 installs
| (0) | Free
Version prompts in Git, run side-by-side across GPT-4/GPT-4o/GPT-4o-mini, compare token cost + latency + output quality, and detect regressions — all inside VS Code.
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

Azure OpenAI PromptOps Manager

Version prompts in Git, run side-by-side across GPT-4 / GPT-4o / GPT-4o-mini, compare token cost + latency + output quality, and detect regressions — all inside VS Code.

Every team using Azure OpenAI manages prompts manually — in Notepad, Slack threads, or random JSON files. There is no versioning, no A/B comparison, no cost-per-prompt tracking, and no regression testing. This extension fixes all of that.


Features

Prompt Library & Versioning

  • Create, edit, and organize prompts in a dedicated .promptops/ directory inside your workspace
  • Version every change with change notes — full Git-friendly history
  • Diff any two versions side by side to see exactly what changed
  • Template variables using {{variable}} syntax with default values
  • Tag-based organization for filtering and grouping prompts
  • Import / Export your entire prompt library as JSON

Side-by-Side Model Comparison

  • Run any prompt across GPT-4, GPT-4o, and GPT-4o-mini simultaneously
  • Rich comparison panel showing output, cost, latency, token usage, and ms/token for each model
  • Visual cost and latency bars for instant readability
  • Best-value highlighting to identify the most cost-efficient model

Token Cost & Latency Tracking

  • Real-time cost calculation per execution using configurable per-model rates
  • Cost Dashboard with 24h / 7-day / all-time summaries
  • Cost breakdown by model and by prompt with visual distribution bars
  • Status bar indicator showing running cost total and execution count

Regression Testing

  • Create baselines from any prompt version on any model
  • Run regressions against new versions to detect output quality changes
  • Jaccard bigram similarity scoring with configurable tolerance thresholds
  • Pass/fail reporting with full baseline vs. test output comparison

Getting Started

1. Install the Extension

Install from the VS Code Marketplace or from the .vsix file.

2. Configure Your Azure OpenAI Endpoint

Run PromptOps: Configure Azure OpenAI Endpoint from the Command Palette (Ctrl+Shift+P) and enter:

  • Your Azure OpenAI endpoint URL (e.g. https://myresource.openai.azure.com)
  • Your API key
  • Deployment names for each model (GPT-4, GPT-4o, GPT-4o-mini)

3. Create Your First Prompt

  1. Click the PromptOps icon in the Activity Bar
  2. Run PromptOps: Create New Prompt
  3. Enter a name, description, system prompt, and user prompt template
  4. Use {{variable}} syntax for dynamic content

4. Run & Compare

  • Single model: PromptOps: Run Prompt
  • Multi-model comparison: PromptOps: Compare Across Models
  • View costs: PromptOps: View Cost Dashboard
  • Regression test: PromptOps: Run Regression Test

Commands

Command Description
PromptOps: Create New Prompt Create a new prompt with system/user templates and variables
PromptOps: Edit Prompt Open a prompt's JSON file for editing
PromptOps: Run Prompt Execute a prompt on a single model
PromptOps: Compare Across Models Run a prompt on multiple models simultaneously
PromptOps: Run Regression Test Test a prompt version against baselines
PromptOps: View Cost Dashboard Open the cost analytics dashboard
PromptOps: Save Prompt Version Save the current prompt state as a new version
PromptOps: Diff Prompt Versions Compare two versions of a prompt
PromptOps: Export Prompt Library Export all prompts and history to JSON
PromptOps: Import Prompts Import prompts from a JSON file
PromptOps: Configure Azure OpenAI Endpoint Set up your endpoint, API key, and deployments
PromptOps: Refresh Prompt Library Refresh the sidebar tree views

Configuration

Setting Default Description
promptOps.endpoint "" Azure OpenAI endpoint URL
promptOps.apiKey "" Azure OpenAI API key
promptOps.deployments {} Model-to-deployment-name mapping
promptOps.apiVersion 2024-06-01 Azure OpenAI API version
promptOps.promptsDir .promptops Directory for prompt storage
promptOps.defaultModel gpt-4o Default model for single runs
promptOps.maxTokens 2048 Default max tokens per completion
promptOps.temperature 0.7 Default temperature
promptOps.costPerMillionInput GPT-4: $30, GPT-4o: $2.50, GPT-4o-mini: $0.15 Input token cost rates
promptOps.costPerMillionOutput GPT-4: $60, GPT-4o: $10, GPT-4o-mini: $0.60 Output token cost rates

Prompt Storage Structure

Prompts are stored in your workspace under .promptops/:

.promptops/
├── prompts/            # Prompt definition JSON files
│   ├── code-review-abc1.json
│   └── summarizer-def2.json
├── history/            # Version history per prompt
│   ├── code-review-abc1.json
│   └── summarizer-def2.json
├── baselines/          # Regression test baselines
│   └── code-review-abc1.json
└── cost-log.json       # Running cost log

All files are plain JSON, fully Git-friendly. Track them in version control to share prompts across your team.


Requirements

  • VS Code 1.85.0 or later
  • Azure OpenAI resource with deployed models (GPT-4, GPT-4o, and/or GPT-4o-mini)
  • A valid API key and endpoint URL

License

MIT

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft