If you don't opt out by Apr 24 GitHub will train on your private repos
What GitHub is changing
- Several commenters clarify the change is about Copilot interaction data, not bulk ingestion of all private repos “at rest.”
- For Free/Pro/Pro+ Copilot users who don’t opt out, GitHub will use:
- Inputs to Copilot (including code snippets and surrounding context),
- Accepted/edited outputs,
- File names, repo structure, navigation, and interaction telemetry.
- Business/Enterprise Copilot customers are said to be excluded from this training change.
- If you don’t use Copilot, multiple comments say this shouldn’t affect you, though some remain skeptical.
Scope and ambiguity
- Many argue the distinction between “private repo data” and “Copilot interaction data” is semantic: Copilot’s “context” effectively includes private repo code.
- The wording of the setting (“Inputs, Outputs, and associated context”) is viewed as vague and legally opaque; even an attorney in the thread finds it hard to interpret.
- It’s unclear how this applies when Copilot is managed by an organization, or when a user is in both personal and enterprise Copilot plans.
Opt-out, UX, and dark patterns
- Major criticism: using opt‑out instead of explicit opt‑in for a new data use.
- Some saw a persistent banner and/or email; others never noticed either, especially those who use GitHub only via CLI.
- The UI copy for the toggle (“you will have access to this feature”) is viewed as misleading, implying you must enable training to “use Copilot.”
- Some note the setting is missing entirely for accounts controlled by orgs or possibly in some regions.
Privacy, legal, and ethical concerns
- Strong distrust of Microsoft/GitHub, framed as “enshittification” and data-grab inevitability once data isn’t end‑to‑end encrypted.
- Multiple mentions of GDPR: opt‑out consent is argued to be invalid; code and commits can contain personal data (names, emails, even health data).
- Worry that contributors using Copilot will leak entire private repos, including sensitive or regulated code (e.g., CUI, secrets).
- Some fear policies can later expand silently; others assume companies may ignore flags in practice with limited recourse.
User responses and alternatives
- Many vow to migrate off GitHub or at least stop using Copilot; some already moved due to the Microsoft acquisition.
- Alternatives mentioned: GitLab, Bitbucket, Sourcehut, Forgejo, Gitea, Codeberg (with caveats), Fossil, self‑hosting on VPS/mini‑PC, and encrypted repos (git-crypt/git-gcrypt).
- A few propose “poisoning” training sets with bad or adversarial code, though others doubt its effectiveness at scale.
- A minority explicitly state they don’t mind training on their code and see benefits (models better matching their style), especially when no secrets are stored in repos.