Skip to main content

GitHub makes Copilot training opt-out for individuals

Starting April 24, GitHub will use Copilot interaction data from Free, Pro, and Pro+ users for model training unless they opt out. Business and Enterprise accounts are excluded.

March 28, 20267 min read
Editorial illustration of a solo developer's Copilot session routing prompts, suggestions, and code context toward model training while a separate enterprise lane stays blocked by contract and policy rails.
ainewssilo.com
GitHub is not just tweaking Copilot's settings. It is changing the default bargain for solo developers.

GitHub is not announcing that it will hoover up every private repository on earth and feed the whole lot into a training run like a Roomba with a law degree. That is the cartoon version, and it is not what the policy says.

The real change is narrower, and more important. Starting April 24, GitHub says it will use Copilot interaction data from Copilot Free, Pro, and Pro+ users to train and improve its AI models unless those users opt out. That interaction data can include prompts, outputs, code snippets, and the surrounding context Copilot sees while you are using it. Copilot Business and Copilot Enterprise are excluded.

That is why this story matters. GitHub is not merely updating a privacy screen. It is changing the default bargain for solo developers. Enterprise customers keep the protection of negotiated contracts. Individual users get a settings toggle, a month of notice, and a polished explanation about industry norms. That may be a rational business decision. It is also the sort of sentence that makes developers reach for the fine print with the enthusiasm of someone checking whether a “free” trial has already found their credit card.

What changed on April 24

GitHub announced the shift on March 25 in both a blog post and a changelog entry tied to updates in its Privacy Statement and Terms of Service. The effective date is April 24. From then on, if you use Copilot Free, Pro, or Pro+ and leave the setting on, GitHub may use your Copilot interactions to develop and train AI models.

GitHub says the setting already exists under Copilot privacy controls, and prior opt-out preferences carry over. If you previously disabled the old prompt-and-suggestion collection setting for product improvements, GitHub says your choice is preserved and your data will not suddenly jump the fence on April 24.

There are two exclusions that matter most. First, Copilot Business and Copilot Enterprise are not included in this training update. Second, GitHub says students and teachers who receive Copilot Pro for free are also not affected.

GitHub says real-world interaction data should make the models better. It also says Microsoft employee interaction data improved acceptance rates in multiple languages. You can believe that and still notice what changed in the bargain. The question is not whether training data helps. The question is who is now supplying it by default.

What code is in scope, and what is not

This is the part that deserves actual reading instead of screenshot-based panic.

GitHub says the data it may collect for training includes accepted or modified outputs, prompts sent to Copilot, code snippets shown to the model, code context around your cursor, comments and documentation you write, file names, repository structure, navigation patterns, and interactions with features such as chat or inline suggestions. In other words, not your entire development life, but certainly more than a single chat prompt and a polite nod.

The company also draws a deliberate line around private repositories at rest. GitHub says it does not use the stored contents of private repositories at rest to train AI models. But it also goes out of its way to explain why it used the phrase “at rest.” If you are actively using Copilot in a private repository, code snippets and associated context from that live session can still be part of the interaction data covered by this policy unless you opt out.

That distinction is subtle, but it is not trivial. “We are not training on your private repo” and “we may train on code Copilot sees while you use it in your private repo” are not the same sentence. One sounds like reassurance. The other is the actual contract.

Editorial diagram showing Copilot prompts, outputs, code snippets, and nearby context flowing from an active coding session toward model training, while private repositories at rest remain separated on a distinct storage lane.
Figure / 01 GitHub's line is specific: active Copilot interactions may train models, but private repositories at rest are not the source pool.

There is another important carve-out in the FAQ. GitHub says it does not train on contents from any paid organization's repositories, even if someone is in that repo with a personal Copilot Free, Pro, or Pro+ subscription. It also says that if a GitHub account is a member of, or outside collaborator with, a paid organization, that user's interaction data is excluded from model training. Yes, the exception has subclauses. Privacy language loves a nesting doll.

GitHub also says the data may be shared with GitHub affiliates, including Microsoft, but not with third-party AI model providers for their own independent training. That last line matters, and it is one of the easier facts to lose once the discourse turns into everyone yelling “private code” at each other across three tabs.

Why developers reacted so hard

The HN and Reddit reaction was not really about discovering that AI companies enjoy more training data. Nobody needed a séance for that. It was about the asymmetry.

Business and Enterprise customers are protected because contracts say so. Individual developers are covered unless they find the setting and switch it off. The trust message there is hard to miss: organizations get negotiated boundaries; solo users get defaults. GitHub can reasonably argue that it is being transparent, and to its credit it did announce the change ahead of time, surface it in email, and preserve earlier opt-outs. But transparency is not the same thing as neutrality, and default settings are where product strategy stops pretending to be philosophy.

That is also why the wording around the toggle annoyed people. In the HN thread, several commenters fixated on the setting being framed like a feature you “have access to,” which landed less like a privacy control and more like a loyalty perk for donating your workflow traces. Nothing says trust quite like a consent screen that sounds as if it was workshopped by a growth team and a tax attorney.

Editorial illustration contrasting a solo Copilot account governed by a settings toggle with a business and enterprise lane protected by contract and organization policy barriers.
Figure / 02 The trust gap is hard to miss. Enterprise customers get contractual exclusion, while individuals get a privacy toggle and some fine print.

I do not think the strongest critique here is “GitHub is evil now.” It is simpler than that. GitHub is redrawing the line between first-party value and user contribution, and the redraw falls hardest on the customers with the least contractual leverage.

What individual developers should do now

If you use Copilot on a personal account and you do not want your interaction data used for model training, go to github.com/settings/copilot/features, open the Privacy section, and disable the training setting before April 24. If you already opted out under the older data-collection control, GitHub says you do not need to do anything.

The broader market lesson is harder to ignore. Coding assistants are no longer just fighting over raw model quality. They are fighting over workflow position, billing boundaries, and access to real developer interaction data. We are seeing the same control instincts elsewhere in the market, whether through the rise of orchestration layers in AI coding's new bottleneck: agent orchestration, or through first-party vendors tightening the edges of third-party access as in Anthropic clamps down on Claude OAuth wrappers. The model is only part of the moat. The default settings are part of it too.

That is why this Copilot change feels bigger than one checkbox. It tells you where the coding-agent market thinks future advantage comes from: not just better weights, but better access to the messy trail real developers leave behind while they work.

If you are fine making that trade, leave the setting on. If you are not, turn it off. But either way, it is worth seeing the change clearly. GitHub did not just improve a product. It changed the deal.

Share this article

Send this story into the feed loop.

Pass the story on without losing the canonical link.

Share to network

Source file

Public source trail

These links anchor the package to the underlying reporting trail. They are not a substitute for judgment, but they do show where the reporting starts.

Primary sourcegithub.blogGitHub Blog
Updates to GitHub Copilot interaction data usage policy

Core announcement covering the April 24 effective date, affected plans, included interaction data, opt-out path, and explicit exclusions.

Primary sourcegithub.blogGitHub Changelog
Updates to our Privacy Statement and Terms of Service: How we use your data

Formal policy and terms framing, including the line between active interaction data and private repositories at rest.

Primary sourcegithub.comGitHub Community
Privacy Statement update on Copilot data use for model training (Free/Pro/Pro+)

FAQ thread with the important carve-outs on students and teachers, enterprise-owned repositories, and paid-organization exclusions.

Backgroundnews.ycombinator.comHacker News
Updates to GitHub Copilot interaction data usage policy

Useful reaction signal showing why the framing and default-toggle language landed badly with many developers. Not a source of record for policy.

Backgroundreddit.comReddit
GitHub to use Copilot data from all user tiers to train AI models

Spillover discussion around opt-out settings and the individual-user impact. Context only.

Backgroundreddit.comReddit
Starting April 24, 2026 GitHub will begin using Copilot interaction data

Preserves the user-facing email copy and the opt-out discussion context. Context only.

Portrait illustration of Talia Reed

About the author

Talia Reed

Staff Writer

View author page

Talia reports on product surfaces, developer tools, platform shifts, category shifts, and the distribution choices that determine whether AI features become durable workflows. She looks for the moment where a launch stops being a demo and becomes an ecosystem move.

Published stories
22
Latest story
Mar 28, 2026
Base
New York

Reporting lens: Distribution is usually the story hiding inside the launch.. Signature: A feature matters when it changes someone else’s roadmap.

Related reads

More AI articles on the same topic.

GitHub makes Copilot training opt-out for individuals | AI News Silo