How we’re responding to The NYT’s data demands in order to protect user privacy
Scope and purpose of the court order
- Many commenters see the order to preserve all ChatGPT logs (including deleted and “temporary” chats) as standard US evidence-preservation practice in a copyright case: NYT wants to quantify how often verbatim or near-verbatim NYT text is generated to calculate damages.
- Others argue this goes far beyond normal proportionality, sweeps in huge amounts of unrelated, highly personal data from uninvolved users, and sets a bad precedent for privacy-focused services.
Privacy, logging, and “legal hold”
- Strong skepticism that OpenAI meaningfully protects privacy: users assume everything sent to a hosted API is logged indefinitely, regardless of marketing claims or toggles.
- Several point out that a “legal hold” is just a preservation requirement; it does not legally block OpenAI from using or accessing the data for other purposes unless other policies/laws do.
- Some say data is a “toxic asset” and the only secure option is not retaining it at all; being forced to keep it inherently increases risk.
Zero Data Retention (ZDR) and product behavior
- Commenters note ZDR APIs exist but are hard to actually obtain; requests are allegedly ignored, leading to accusations that ZDR is more marketing than reality.
- OpenAI’s own post says ZDR API endpoints and Enterprise are excluded from the order, but people question why privacy is a paid/approved feature rather than a universal option.
- There is confusion and criticism around the in-app “Improve the model for everyone” toggle versus the separate privacy portal, seen by some as a dark pattern.
GDPR and non-US users
- Debate over whether complying with the US order violates GDPR:
- Some say GDPR has allowances for court-ordered retention and it’s only a problem if data is kept beyond the case.
- Others cite GDPR limits on honoring third-country orders without specific agreements and argue an EU court might bar such retention for EU residents.
NYT vs OpenAI copyright dispute
- Several think NYT’s underlying claim is strong, pointing to examples where ChatGPT allegedly regurgitates NYT text and arguing per-infringement damages justify broad discovery.
- Others view OpenAI’s training as fair use and call NYT’s demand overbroad or abusive of US discovery rules.
- OpenAI’s public framing of the lawsuit as “baseless” and as a privacy attack is widely characterized as spin; critics say OpenAI’s own copyright decisions created this situation.
Government and surveillance concerns
- A long subthread debates whether US intelligence agencies likely access such data:
- Some assert it’s almost certainly tapped and easily searchable using modern methods.
- Others call this unfalsifiable conspiracy thinking, noting legal and technical barriers, but still concede metadata alone is highly revealing.
Sensitivity of LLM chat histories
- Many emphasize that LLM conversations can be more revealing than browser history: people use them for emotional processing, relationship issues, work drafts, and “raw” inner thoughts, making the retention order feel especially invasive.