Discord just killed anonymity
Scope of the Discord change (what’s actually gated)
- Several commenters note the article’s headline is misleading: age verification is not required to create an account, join servers, text chat, or use normal voice channels.
- Verification (or “age assurance”) is required for:
- Viewing unblurred “sensitive” / NSFW content and disabling filters
- Entering age-gated channels/servers/commands
- Changing DM/friend-request safety settings
- Speaking in Stage channels (but not in regular voice channels)
- Some users already see messages hidden behind prompts that require ID to view, making the restriction feel effectively mandatory for many community contexts.
Anonymity and community impact
- Many argue anonymity on Discord was already weak: IP logging, email, phone-number enforcement, VPN-fingerprinting, and admin tools made it unsuitable for serious anonymity.
- Still, users say this step will destroy or shrink niche, sensitive, or stigmatized communities (NSFW, LGBT, politics, etc.) whose members won’t “doxx themselves” to Discord.
- Others counter that for casual gaming chats nothing changes, and the majority will simply not verify and keep using Discord.
Privacy, surveillance, and age-inference concerns
- Discord says: facial scans stay on-device, IDs are used only to derive age then deleted, and an internal ML model infers age groups from behavioral signals (servers, activity patterns, etc.), without reading message content.
- Commenters worry about:
- False classifications (e.g., minors flagged as adults)
- Pressure to upload government ID or face scans to third-party vendors
- Long-term risks of leaks, data brokerage, and doxxing tied to real-world identity and sensitive content.
- Some see this as part of a broader “surveillance state / SEXINT” trajectory; others dismiss that as conspiratorial.
Law, liability, and enshitification
- Several point to UK-style online safety laws requiring “highly effective” age verification/estimation; Discord may be preemptively complying.
- Others frame it as classic pre-IPO “enshitification” and reputational risk management: be “safe for teens,” appease regulators, payment processors, and investors.
User behavior and alternatives
- Many predict only a small privacy-conscious minority will leave; network effects dominate, as seen with Reddit/Netflix controversies.
- Others report already moving to IRC, Matrix, Steam, or self-hosted tools, accepting fewer features for more control and less platform-level policing.
- There’s debate whether self-hosting or decentralized tools truly improve anonymity versus just shifting who controls the data.