Fine-tuned models are now in limited public beta for GitHub Copilot Enterprise


Our mission is to create an AI assistant that’s not only ubiquitous and trustworthy but also uniquely tailored to each organization’s needs. By integrating features like repository indexing and knowledge bases, we’ve enhanced GitHub Copilot to deliver more contextual and personalized assistance.

Now, we’re taking this mission a step further with the introduction of fine-tuned models, available in public preview for Copilot Enterprise users today. This new capability lets enterprise users customize the Copilot with their proprietary codebases and coding practices, providing increased relevance, quality, and consistent code completion support that’s aligned with their specific needs.

The power of a fine-tuned GitHub Copilot

GitHub Copilot is already making a significant impact on how developers and organizations build software, with impressive results. Our research with Accenture shows that developers using Copilot experienced an 8% increase in pull requests, a 15% boost in pull request merge rates, and an 84% increase in build success rate. The research also found that 90% of developers were more fulfilled in their job when using GitHub Copilot, and 95% said they enjoyed coding more with the help of Copilot. These results highlight the effectiveness Copilot has in driving productivity, code quality, and developer happiness across organizations.

Fine-tuned models represent the next big leap in customization, now extending these capabilities directly into the in-line code completion experience. By training Copilot on your private codebases and also incorporating telemetry from how your team interacts with Copilot suggestions, fine-tuned models enable Copilot to adapt to your organization’s unique coding environment with latency that works seamlessly in real-time coding. Copilot becomes intimately familiar with your modules, functions, rare languages like legacy or proprietary languages, and internal libraries—delivering code suggestions that are not just syntactically correct, but more deeply aligned with your team’s coding style and standards.

Customize GitHub Copilot for your needs

Any organization that relies on internal APIs, specialized frameworks, proprietary languages, or strict coding styles can benefit from fine-tuned models. For example, finance organizations that work with legacy languages like COBOL can leverage fine-tuned models to address their specific coding requirements. On the other hand, sectors such as technology or healthcare, which often rely on internal libraries to enforce compliance and security standards—like ensuring cloud resources are deployed in line with organizational policies—can see meaningful improvements in coding accuracy and efficiency. Whether your team is operating in these industries or others, fine-tuned models help Copilot become a powerful, customized tool tailored to your unique needs.

With fine-tuned models, your developers receive code that requires fewer adjustments, which can help new team members onboard more quickly and experienced developers focus more on creating rather than correcting. This is a major step forward in making Copilot an even more helpful tool for your organizations, by providing more relevant, high-quality, and consistent coding support.

How we fine-tune custom models for GitHub Copilot

It’s true that we’ve been on a customization journey with GitHub Copilot, and we know that how we implement customization really matters. Copilot Enterprise first received repository indexing and knowledge bases—both powered by retrieval-augmented generation (RAG), a lightweight retrieval method used to gather information from sources outside of the model’s original training data without needing to retrain the entire model. While RAG has been effective for enhancing the chat experience with up-to-date outputs, it doesn’t meet the performance demands required for real-time code completion. With fine-tuning, we’re bringing customization to the code completion experience for the first time, allowing Copilot to deliver contextualized suggestions with the speed necessary for inline coding.

We customize each model using the LoRA (Low-Rank Approximation) method, which fine-tunes a smaller subset of the most important model parameters during the supervised training phase, making the model more manageable and efficient. For users, this means faster and more affordable training compared to traditional fine-tuning techniques. Additionally, the fine-tuning process incorporates insights from how your team interacts with suggestions from Copilot, ensuring that the model aligns more closely with your organization’s specific needs.

Every time you train a custom model, it starts with our base model for in-line code completion. As we upgrade our base models, future custom model retrainings will automatically roll forward to these new models as soon as fine-tuning is available. The fine-tuning process is powered by the Azure OpenAI Service, which provides scalability and security throughout the training pipeline.

Security and privacy in fine-tuned models

We know privacy and security are paramount, so we’ve taken critical steps to incorporate data security steps for fine-tuned models at the scale that enterprises need. Importantly, your data is always yours. It is never used to train another customer’s model, and your custom model remains private, giving you full control.

Image of a flowchart showcasing the high-level technical architecture of GitHub Copilot fine-tuned models, with specific callouts for what is client managed, GitHub managed, and Azure managed.

When you initiate a training process, your repository data and telemetry data are tokenized and temporarily copied to the Azure training pipeline. Some of this data is used for training, while another set is reserved for validation and quality assessment. Once the fine-tuning process is complete, the model undergoes a series of quality evaluations to ensure it outperforms the baseline model. This includes testing against your validation data to confirm that the new model will improve code suggestions specific to your repositories.

If your model passes these quality checks, it is deployed to a Multi-LoRA Azure deployment. This setup allows us to host multiple LoRA models at scale while keeping them network isolated from one another. After the process is complete, your temporary training data is removed from all surfaces, and data flow resumes through the normal inference channels. The Copilot proxy services ensure that the correct custom model is used for your developers’ code completions.

How to participate in the limited public beta

Fine-tuned models are currently in limited public beta, and we are gradually onboarding customers from our waitlist. If you’re already on the waitlist, we sincerely appreciate your patience as we proceed with the onboarding process. If you haven’t joined the waitlist yet, you can do so by clicking here.

Written by

Blog Article: Here

  • Related Posts

    National Robotics Week — Latest Physical AI Research, Breakthroughs and Resources

    This National Robotics Week, running through April 12, NVIDIA is highlighting the pioneering technologies that are shaping the future of intelligent machines and driving progress across manufacturing, healthcare, logistics and more. Check back here throughout the week to learn the latest on physical AI, which enables machines to perceive, plan and act with greater autonomy
    Read Article

    Celebrating Microsoft’s 50 years

    Satya Nadella, Chairman and CEO, shared the following remarks at Microsoft’s 50th anniversary today. It’s so wonderful to be here with all of you celebrating 50 years of Microsoft. And it’s especially exciting to be doing it at a time like this. For me, though, it starts with Bill [Gates] and Steve [Ballmer], who are…

    The post Celebrating Microsoft’s 50 years appeared first on The Official Microsoft Blog.

    Leave a Reply

    Your email address will not be published. Required fields are marked *

    You Missed

    Amazon Nova Reel 1.1: Featuring up to 2-minutes multi-shot videos

    Amazon Nova Reel 1.1: Featuring up to 2-minutes multi-shot videos
    AWS Weekly Review: Amazon EKS, Amazon OpenSearch, Amazon API Gateway, and more (April 7, 2025)

    The Changing Role of Developers in the Age of AI Agents

    The Changing Role of Developers in the Age of AI Agents

    “Bazooka Innovation” for Professional Services: How Certinia and Agentforce are Driving Customer Success

    “Bazooka Innovation” for Professional Services: How Certinia and Agentforce are Driving Customer Success

    Bringing multimodal search to AI Mode

    Bringing multimodal search to AI Mode

    New AI-powered experiments from Google Arts & Culture Artists in Residence

    New AI-powered experiments from Google Arts & Culture Artists in Residence