Skip to content
| Marketplace
Sign in
Visual Studio Code>SCM Providers>PR Merge Conflict ResolverNew to Visual Studio Code? Get it now.
PR Merge Conflict Resolver

PR Merge Conflict Resolver

codextio

|
72 installs
| (0) | Free
Visualize, resolve, and sync merge conflicts for Pull Requests from Azure DevOps or GitHub directly within VS Code.
Installation
Launch VS Code Quick Open (Ctrl+P), paste the following command, and press enter.
Copied to clipboard
More Info

PR Merge Conflict Resolver

Resolve Azure DevOps and GitHub Pull Request merge conflicts directly inside Visual Studio Code with a clean, side-by-side diff editor, live conflict tracking, and AI-powered conflict resolution.

No browser switching. No manual rebasing.
Resolve, review, and submit merges where you write code.


✨ Key Features

  • 🔗 Connect to Azure DevOps & GitHub Git repositories
  • 🔐 Authenticate via Personal Access Token (PAT) or OAuth 2.0
  • 📂 Browse Projects → Repositories → Pull Requests
  • ⚠️ Automatically detect merge conflicts in PRs
  • 🧩 Resolve conflicts using a side-by-side diff editor
  • 🎯 Accept changes from Source or Target branch
  • 📁 Bulk Resolve - Accept all source/target for entire folders, groups, or all pending files
  • 🤖 AI-Powered Resolution with multi-LLM support (OpenAI, Azure AI, Gemini, Claude, and more)
  • � Semantic Merge Pipeline - Chunks files semantically and merges differing blocks via LLM (NEW in v1.1.0)
  • 🔀 Smart Signature Merging - Automatically creates UNION of method parameters from both branches (NEW in v1.1.0)
  • 📝 Ignore Formatting - Option to hide whitespace/indentation noise in diffs (NEW in v1.1.0)
  • �🧠 LLM Embeddings & Semantic Search with RAG orchestration for intelligent context retrieval
  • 📊 Real-time statistics showing conflict resolution progress
  • 🚀 Submit resolved merges back to the PR from VS Code

🧠 LLM Embeddings & Semantic Search (NEW)

Advanced RAG-Based Conflict Resolution

The extension now features a sophisticated LLM Embedding Service that uses semantic search with multi-dimensional vectors for intelligent context retrieval:

Feature Description
Semantic Search Uses cosine similarity to find the most relevant code sections
Multi-Dimensional Vectors Support for 768D to 4096D embeddings depending on provider
Intelligent Code Chunking Automatically detects imports, classes, methods, and comments
Method-Level Analysis Embeds individual method conflicts for targeted resolution
Preference Embedding Converts your merge preferences into vector space for better context
RAG Orchestration Combines semantic search results with user preferences for optimal prompts

Supported Embedding Providers

Provider Embedding Model Dimensions Notes
OpenAI text-embedding-ada-002 1536D Best quality, requires API key
Azure OpenAI text-embedding-ada-002 1536D Enterprise deployment
Google text-embedding-gecko 768D Good performance
Ollama nomic-embed-text 4096D Local, no API key needed
Cohere embed-english-v3.0 1024D Multilingual support
Fallback TF-IDF 1536D Works offline when APIs unavailable

How Semantic Search Works

  1. Content Indexing: Source and target code are chunked intelligently by structure (imports, classes, methods)
  2. Embedding Generation: Each chunk is converted to a high-dimensional vector using your configured LLM provider
  3. Preference Embedding: Your AI merge preferences are also embedded for context matching
  4. Similarity Search: Cosine similarity finds the most relevant code sections for the conflict
  5. RAG Context Building: Relevant chunks are assembled with similarity scores into the prompt
  6. Enhanced Resolution: The LLM receives rich, contextually relevant information for better merges

🧬 Semantic Merge Pipeline (NEW in v1.1.0)

The AI merge engine now uses a semantic merge pipeline as the primary approach, delivering significantly more accurate results:

How It Works

  1. Semantic Chunking: Both source and target files are parsed into meaningful blocks — imports, classes, methods, comments
  2. Cosine Similarity Matching: Each source chunk is matched to its target counterpart using vector similarity
  3. Selective LLM Merge: Only chunks that actually differ are sent to the LLM — reducing tokens and improving accuracy
  4. Source Fingerprint Verification: After merging, unique source "fingerprints" (new params, method calls, strings) are verified in the output
  5. Automatic Patching: If source changes were dropped, the pipeline automatically patches them back in
  6. Signature Enforcement: A final pass ensures all method parameters from both branches are present in the merged output

Smart Method Signature Merging

Scenario Result
Source adds params Uses SOURCE signature (enhancement)
Target adds params Creates UNION with ALL params from BOTH branches
Both add different params Combines ALL unique params from both sides
Same signature, body changed Auto-applies source body changes
New method in source Auto-inserted into merged output
Method removed in source Auto-removed from merged output

Example:

Source: CreateOrder(int orderId, bool validateInventory)
Target: CreateOrder(int orderId)
Result: CreateOrder(int orderId, bool validateInventory)  // Union of all params

🤖 AI-Powered Conflict Resolution

Multi-LLM Provider Support

The extension supports 7 LLM providers for intelligent conflict resolution:

Provider Models Features
Azure AI GPT-5, GPT-4.1, GPT-4o, o3 Enterprise-ready, embeddings support
OpenAI GPT-5.2 Pro, GPT-5 Codex, GPT-4.1 Best coding models, Codex series
Google Gemini Gemini 3 Pro, 2.5 Flash 1M context window, free tier
Anthropic Claude Claude Opus 4.5, Claude Sonnet 4.5 Advanced reasoning
Ollama Llama 3.3, CodeLlama, Mistral Local/private, no API key
Moonshot Kimi Kimi K2, Moonshot V1 128K context
OpenRouter All providers via single API Unified access

How AI Resolution Works

  1. Select a conflicted file in the merge editor
  2. Click "Resolve with AI" button
  3. The AI analyzes both source and target changes with full context
  4. Receive an intelligent merged resolution
  5. Review, edit if needed, and accept

🚀 Getting Started

1. Install the Extension

Search for "PR Merge Conflict Resolver" in the VS Code Extensions Marketplace and install it.

2. Open the Extension

Click the PR Merge Conflict Resolver icon in the Activity Bar or use the Command Palette (Ctrl+Shift+P):

PR Merge Conflict: Configure PR Provider

Activity Bar Icon

3. Configure Git Provider

  • Click on gear icon of Repositories section or Ctrl+Shift+P and select PR Merge Conflict: Configure PR Provider
  • Choose between Azure DevOps or GitHub as your Git provider

Configure Git Provider Icon

Azure DevOps

When first launched, you'll be prompted to configure:

  • Azure DevOps Organization name
  • Personal Access Token (PAT) for authentication
  • Project - selected from a list or entered manually
  • Repository to work with

Project Selection Options:

After entering your organization and PAT, you'll be prompted to choose how to select the project:

Option Description When to Use
📋 Select from Project List Shows a dropdown of all available projects When your PAT has permission to list all projects in the organization
✏️ Enter Project Name Manually Type the project name directly When you only have access to specific projects (no org-wide project list permission)

Tip: If fetching the project list fails, you'll automatically be offered the option to enter the project name manually. This ensures the extension works for users with any level of access permissions.

Extension Overview

GitHub

For GitHub repositories:

  • GitHub Personal Access Token with repo scope
  • Owner/Organization name
  • Repository selection Extension Overview

Configure Azure DevOps

4. Select a Pull Request

  • Browse available Pull Requests in the selected repository
  • PRs with merge conflicts are automatically highlighted
  • Select a PR to view its conflicted files

Select Pull Request

5. Resolve Conflicts

  • Click on any conflicted file to open the merge editor
  • View source and target branches side-by-side
  • Accept changes from either branch or edit manually
  • Save your resolution

Resolve Conflicts - Side by Side View

Edit using quick actions: you can edit manually or use the Accept from Source/Target buttons to quickly resolve sections.

Edit Manually

6. Submit Changes To PR

  • Once all conflicts in a file are resolved, click Submit Merge to PR
  • Enter a commit message
  • Changes are pushed directly back to your Azure DevOps PR

Submit Merge


📁 Bulk Resolve Conflicts (NEW)

Quickly resolve multiple conflict files at once using the context menu:

How to Use

  1. Right-click on any of these items in the Conflict Files tree:

    • Folder - Resolves all conflict files within that folder
    • Filter Group (Source Branch Files / Other PR Files) - Resolves all files in that group
    • Section Header (Pending Conflicts) - Resolves ALL pending files
  2. Select "Accept All Source" or "Accept All Target"

  3. Files are resolved locally with the chosen branch content

  4. Click "Sync Resolved Files" (cloud icon) in the toolbar to push changes

Features

Feature Description
Batch Resolution Resolve entire folders or all pending files with one click
Smart Sync Button Only appears when there are locally resolved files ready to push
Per-File Progress Shows progress for each file during sync (e.g., "file.ts (3/10)")
Error Recovery Files that fail to sync are automatically marked as pending again
Resolution Tracking Commit messages show how each file was resolved (source/target)

🤖 Configure LLM for AI-Powered Resolution

To use AI-powered conflict resolution, you need to configure an LLM provider:

1. Open LLM Configuration

Use the Command Palette (Ctrl+Shift+P) and run:

PR Merge Conflict: Configure LLM

2. Select LLM Provider

Choose from the available providers:

  • Azure AI - Enterprise Azure OpenAI deployments
  • OpenAI - Direct OpenAI API access
  • Google Gemini - Google's AI with large context windows
  • Anthropic Claude - Advanced reasoning capabilities
  • Ollama - Local models (free, no API key needed)
  • Moonshot Kimi - Large context Chinese/English models
  • OpenRouter - Access multiple providers with one API key

LLM Configuration

3. Enter API Key

  • For cloud providers (OpenAI, Azure, Gemini, Anthropic, etc.), enter your API key
  • For Ollama, no API key is required - just ensure Ollama is running locally

4. Configure Endpoint (If Required)

  • Azure AI: Enter your Azure OpenAI endpoint (e.g., https://your-resource.openai.azure.com)
  • Ollama: Enter your local endpoint (default: http://localhost:11434)

5. Select Models

  • Choose the chat model for conflict resolution
  • Optionally select an embedding model for enhanced context understanding

6. Verify Configuration

Once configured, you'll see a confirmation message. You can now use the "Resolve with AI" button on any conflicted file.

Resolve with AI Button Click

7. AI Starts Resolving

Clicking the button sends the conflict data to the selected LLM, which processes it and returns a merged resolution. During this process, you'll see a loading indicator and see cancel button if you wish to abort.

LLM Resolving Conflict

7. AI-Powered Conflict Resolution

The AI suggests a merged resolution based on the context from both branches. You can review the file changes and Submit Merge To PR.

LLM Resolution

Tip: For best results with code conflicts, use models optimized for coding like Gemini 3 Pro, Claude Opus 4.5, Claude Sonnet 4.5, GPT-5.2 Pro, or KIMI K2.


🧭 Workflow Overview

  1. Connect to Azure DevOps with your PAT
  2. Select organization, project, and repository
  3. Choose a Pull Request with conflicts
  4. View conflicted files in the Conflict Files panel
  5. Resolve conflicts using the side-by-side editor
  6. Track progress in the Conflict Statistics panel
  7. Submit resolved files back to the PR

🔐 Authentication

The extension supports multiple authentication methods:

Azure DevOps

Personal Access Token (PAT) - Recommended

  1. Go to Azure DevOps → User Settings → Personal Access Tokens
  2. Create a new token with:
    • Code (Read & Write) permissions
    • Pull Request Threads (Read & Write) permissions
  3. Copy the token and paste it when prompted by the extension

OAuth 2.0

OAuth authentication is also supported for seamless integration with your Azure DevOps account.

GitHub

Personal Access Token

  1. Go to GitHub → Settings → Developer Settings → Personal Access Tokens
  2. Generate a new token (classic) with:
    • repo scope (Full control of private repositories)
  3. Copy the token and use it in the extension configuration

Security: All credentials are stored securely using VS Code's Secret Storage API. Tokens are never logged or transmitted outside of Azure DevOps/GitHub API calls.


📊 Features in Detail

Conflict Detection

  • Automatically scans PRs for merge conflicts
  • Displays conflicted files grouped by status (Pending/Resolved)
  • Shows real-time statistics on conflict resolution progress

Side-by-Side Merge Editor

  • Left panel: Source branch changes
  • Right panel: Target branch / merged view
  • Quick actions to accept from source or target
  • Manual inline editing for complex conflicts

Progress Tracking

  • Visual indicators for pending and resolved conflicts
  • Live statistics panel showing:
    • Total conflicts count
    • Pending conflicts
    • Resolved files
    • Overall resolution percentage

Statistics Panel

Direct PR Integration

  • Submit resolved files directly to the PR
  • Add commit messages for each resolution
  • No need to switch to browser or use git commands

🛠 Requirements

  • Visual Studio Code version 1.85.0 or higher
  • Azure DevOps or GitHub account with access to repositories
  • Personal Access Token with appropriate permissions:
    • Azure DevOps: Code (Read & Write), Pull Request Threads (Read & Write)
    • GitHub: repo scope
  • LLM API Key (optional, for AI-powered resolution):
    • Microsoft Foundry (Azure AI), OpenAI, Gemini, Anthropic, etc.
    • Or use Ollama for local, free AI resolution

📦 Extension Settings

Configure the extension through VS Code settings:

Provider Settings

  • prMergeConflictResolver.provider: Git provider (azureDevOps or github)
  • prMergeConflictResolver.azureDevOps.organization: Azure DevOps organization name
  • prMergeConflictResolver.azureDevOps.project: Azure DevOps project name
  • prMergeConflictResolver.azureDevOps.repository: Azure DevOps repository name
  • prMergeConflictResolver.github.owner: GitHub owner/organization name
  • prMergeConflictResolver.github.repository: GitHub repository name

LLM Settings

  • prMergeConflictResolver.llm.provider: LLM provider (azure, openai, gemini, anthropic, grok, deepseek, ollama, kimi, openrouter)
  • prMergeConflictResolver.llm.model: Model ID for the selected provider
  • prMergeConflictResolver.llm.embeddingModel: Embedding model for RAG-based resolution

Diff Settings

  • prMergeConflictResolver.diffContext: Number of context lines in diff (default: 3)
  • prMergeConflictResolver.autoRefreshInterval: Auto-refresh interval in seconds (0 to disable)

🐛 Troubleshooting

"Failed to fetch repositories"

  • Verify your PAT has the correct permissions
  • Check that your organization name is correct
  • Ensure you have access to the specified project

"Command not found" errors

  • Reload VS Code window (Ctrl+Shift+P → "Reload Window")
  • Check the extension is activated in the Extensions panel

Conflicts not showing

  • Click the refresh icon in the Pull Requests panel
  • Verify the PR actually has merge conflicts in Azure DevOps

📄 License

This extension is licensed under the MIT License.


🤝 Support

For issues, feature requests, or contributions:

  • Contact: codextio.dev@gmail.com

🎯 Roadmap

Features completed:

  • ✅ Azure DevOps support
  • ✅ GitHub support
  • ✅ Side-by-side merge editor
  • ✅ Real-time conflict statistics
  • ✅ Multi-LLM provider support (7 providers)
  • ✅ AI-powered conflict resolution
  • ✅ RAG-based context understanding (in memory limits)
  • ✅ LLM Embeddings with semantic search (cosine similarity)
  • ✅ Multi-dimensional vector support (768D - 4096D)
  • ✅ Intelligent code chunking (imports, classes, methods)
  • ✅ Preference embedding for personalized resolution
  • ✅ Semantic merge pipeline with chunk-level LLM merging
  • ✅ Smart method signature merging (parameter union)
  • ✅ Bulk conflict resolution
  • ✅ Ignore formatting option for diff view
  • ✅ Pipeline details modal with execution statistics
  • ✅ Premium/Regular model tier optimization for predictable AI behavior
  • ✅ Custom temperature manual entry for fine-tuned control

Future enhancements planned:

  • Advanced diff algorithms
  • Conflict resolution templates
  • Team collaboration features
  • Multi-file context awareness

📋 Release Notes

v1.1.1 (February 2026 - Patch Release)

🎯 Model Tier Optimization & Keybinding Fix

  • NEW: Premium vs Regular model tier selection — optimizes prompts, batch sizes, and post-processing per model capability for more predictable and accurate AI resolutions
  • NEW: Custom temperature manual entry — enter any value within provider-specific range alongside preset options (Precise, Balanced, Creative)
  • NEW: Dedicated Ctrl+G Go to Line shortcut — exclusively opens the extension's Go to Line modal when the Merge Conflict Resolver webview is focused

v1.1.0 (February 2026 - Minor Feature Release)

🧬 Semantic Merge Pipeline & Smart Signature Merging

  • NEW: Semantic merge pipeline — files chunked into semantic blocks, matched via cosine similarity, merged individually via LLM
  • NEW: Source-content fingerprint verification ensures source changes are never dropped
  • NEW: Signature enforcement pass guarantees union of all parameters from both branches
  • NEW: Ignore Formatting option in diff view to hide whitespace/indentation noise
  • NEW: Pipeline Details Modal with per-stage token/time breakdown and execution statistics
  • NEW: Token usage display in AI Summary modal showing total tokens consumed
  • IMPROVED: Smart method signature merging creates UNION of parameters instead of dropping them
  • IMPROVED: LLM resolution strategy optimized for thinking models with batched conflict resolution
  • IMPROVED: Pipeline timeouts increased for complex resolutions (60s architect, 90s builder)
  • IMPROVED: Progress notification messages now sync with progress modal steps
  • IMPROVED: Method overload detection using methodName__paramCount keying
  • FIXED: Whitespace normalization, diff computation, scroll position, search accuracy
  • FIXED: Success modal display and progress modal state transitions

v1.0.26 (February 2026)

📁 Bulk Resolve Conflicts

  • NEW: Right-click on folders, filter groups, or section headers to resolve all files at once
  • NEW: Smart Sync button only appears when there are locally resolved files ready to push
  • NEW: Resolution type tracking for proper commit messages
  • FIXED: Inline button click issues and context menu commands

v1.0.25 (February 2026)

🔍 Enhanced Search Functionality

  • NEW: In-panel search modal with Ctrl+F shortcut for both Source and Target views
  • NEW: Real-time search highlighting with match navigation (Enter/Shift+Enter)
  • NEW: Search results counter showing current position (e.g., "3/10")
  • NEW: Backdrop overlay technique for persistent highlights in edit mode
  • IMPROVED: Pixel-perfect scroll positioning using actual element positions
  • IMPROVED: Focus management keeps search input focused during navigation
  • FIXED: Search now accurately scrolls to matches deep in large files
  • FIXED: Highlights remain visible in textarea edit mode
  • FIXED: "Accept Source" button now properly updates target view with source content
  • FIXED: Scroll position preserved when toggling edit mode in target view
  • FIXED: LLM status label centered in header area

v1.0.24 (February 2026)

✨ AI Summary & UI Improvements

  • NEW: Enhanced AI Summary Modal with two sections:
    • "Conflict Resolutions" - shows actual code changes (statements, methods, line changes)
    • "Preferences Resolutions" - shows method preference decisions
  • NEW: Resolution status labels with color coding ("Used Source", "Kept Target", "Merged")
  • NEW: AI Summary now displays for all successful resolution scenarios (Stage 1, Stage 2, Programmatic Fix)
  • IMPROVED: Green highlighting for all resolved conflicts in target view
  • IMPROVED: Local Analysis modal with green radio buttons and conditional Apply button
  • FIXED: Template rendering issues with Handlebars syntax
  • FIXED: Edit icon display in toolbar
  • CLEANED: Removed debug console.log statements

v1.0.23 (February 2026)

🧠 LLM Embeddings & Semantic Search

  • NEW: LLM Embedding Service with multi-provider support
  • NEW: Semantic search using cosine similarity for intelligent context retrieval
  • NEW: Multi-dimensional vector support (768D - 4096D depending on provider)
  • NEW: Intelligent code chunking that detects imports, classes, methods, and comments
  • NEW: Method-level conflict embedding for targeted resolution
  • NEW: User preference embedding for personalized AI responses
  • NEW: RAG orchestration combining semantic search with user preferences
  • NEW: Automatic fallback to TF-IDF when LLM embedding APIs are unavailable
  • IMPROVED: Enhanced AI resolution prompts with similarity scores and context

v1.0.22 (January 2026)

🤖 AI Merge Preferences & Local Analysis Improvements

  • NEW: AI Merge Preferences form with 9 configurable options
  • NEW: Method-level conflict decisions (source/target/AI merge)
  • NEW: Special instructions support for custom merge behavior
  • NEW: Open-Router LLM provider with manual model option
  • NEW: Local Analysis modal with improved UI
  • NEW: Method/function signature conflict detection support
  • Added vertical scrollbars to source/target code blocks
  • Reduced stat card heights in Local Analysis Summary View

v1.0.21 (January 2026)

✨ Enhanced Azure DevOps Configuration

  • NEW: Manual project name entry option for users with restricted project list permissions
  • IMPROVED: Automatic fallback to manual entry when project list fetch fails
  • IMPROVED: Flexible project selection workflow accommodating various permission levels 🐛 Bug Fixes
  • FIXED: Removed file in Pull Request now handled correctly and resolved without errors

v1.0.20 (January 2026)

🚀 Initial Release 📊 Local Analysis

  • Added local conflict analysis without LLM requirement
  • Statement-level change detection
  • Source/target code comparison view
  • Summary statistics for local analysis

v1.0.19 (December 2025)

🤖 Multi-LLM Provider Support

  • Added 7 LLM providers (Azure AI, OpenAI, Google Gemini, Anthropic Claude, Ollama, Moonshot Kimi, OpenRouter)

v1.0.0 (December 2025)

  • Azure DevOps integration
  • GitHub integration
  • Side-by-side merge editor
  • Real-time conflict statistics
  • Direct PR submission
  • PAT and OAuth authentication

⚡ Powered by Codextio

  • Contact us
  • Jobs
  • Privacy
  • Manage cookies
  • Terms of use
  • Trademarks
© 2026 Microsoft