Open Source AI Editor: Second Milestone
November 6th, 2025 by the VS Code Team
In May, we announced our initial plan to make VS Code an open source AI editor, and in June, we reached our first milestone by open sourcing the GitHub Copilot Chat extension.
While chat was a significant step forward, an important part of our AI functionality still remained: the inline suggestions that appear as you type. Today, we're reaching that next milestone in our journey: inline suggestions are now open source.

One extension, same user experience
For the past few years, GitHub Copilot in VS Code has been split across two extensions: the GitHub Copilot extension (for ghost text suggestions) and the GitHub Copilot Chat extension (for chat and next edit suggestions). We are working towards providing all Copilot functionality in a single VS Code extension: Copilot Chat.
To achieve this, we are now testing disabling the Copilot extension and serving all inline suggestions from Copilot Chat. We have ported the vast majority of features into the chat extension, so the progressive rollout of a single extension experience should feel consistent and transparent to everyone.
Nothing should change in your experience. You'll continue to get the same intelligent code suggestions as you type, plus all the chat and agent mode features you're already using. If you encounter any problems, please report an issue or see how to use the previous experience if needed.
As part of this refactoring, the GitHub Copilot extension will be deprecated by early 2026, which means it will be removed from the VS Code Marketplace.
We've also simplified our terminology: we now use inline suggestions to refer to all AI-generated code suggestions that appear as you type (including ghost text and next edit suggestions). We continue working to unify the actual product experiences as well, including the UX and timing for different kinds of suggestions.
Explore and contribute
With inline suggestions available in the vscode-copilot-chat repository, you can explore and contribute to how they work:

- "Typing-as-suggested" detection - As you type, the extension first checks if you're following a previous suggestion and can continue showing it without making a new request
- Caching - If not typing as suggested, the extension checks if cached suggestions can be reused to improve performance
- Reusing ongoing requests - If no cached suggestions are available, the extension checks if there's an ongoing LLM request from the previous keystroke that hasn't finished streaming back yet. Since this ongoing request is likely similar to the current request, the extension reuses it instead of firing off a new request and canceling the ongoing one, which significantly improves performance
- Prompt construction - If no ongoing request can be reused, the extension gathers relevant context from your current file, open files, and workspace, then formats it into a prompt to send to the LLM
- Model inference - The extension requests inline suggestions from multiple providers: ghost text suggestions for the current cursor position, and next edit suggestions that predict where you might edit next. Ghost text suggestions at the cursor are prioritized when available; otherwise, next edit suggestions are used
- Post-processing - Raw model outputs are refined to ensure they fit your code style, indentation, and syntax
- Multi-line intelligence - The extension decides whether to show a single line or multiple lines, based on confidence and context
Performance improvements
Along with consolidating into a single extension, this refactoring has led to technical improvements to inline suggestions:
- Reduced latency - We fixed networking issues to optimize how suggestions are delivered, enabling the chat extension to serve ghost text faster
- Quality validation - We ran extensive experiments to ensure there are no regressions in either latency or suggestion quality
Troubleshooting
As with all changes, despite our best efforts, there is a chance that we missed something! If you encounter any issues with the unified extension experience, you can temporarily revert to the previous two-extension behavior by unchecking the unification setting:

What's next?
The next phase of our OSS journey is to refactor some AI features and components from the Copilot Chat extension into VS Code core. We're excited to continue this journey with the community and shape the future of development as an open source AI editor.
We'll continue actively improving our inline suggestions experiences - as always, you can follow along on our iteration plans for the latest:

We welcome your feedback and contributions. Feel free to open pull requests and file issues.
Happy coding! 💙
The VS Code Team