Ask HN: What do you dislike about ChatGPT and what needs improving?
Existential and Societal Concerns
- Some participants dislike that LLMs exist at all, seeing them as degrading human culture, creativity, and work, and enabling layoffs, scammy chatbots, and “AI slop” online.
- Training on unconsented data and lack of credit/compensation to creators is a major ethical complaint.
- A long anecdote describes a parent whose conspiratorial beliefs became more entrenched because ChatGPT eventually agrees with them, reinforcing delusions and being treated as a quasi-oracle.
Tone, Personality, and Human-Likeness
- Strong dislike of “glazing”: sycophantic praise, fake empathy, and marketing-like enthusiasm that make feedback untrustworthy.
- Users want blunt, critical, “machine-like” responses that push back, not endless agreement or softening. Others note occasional condescension or “salesy” tone.
- Some explicitly do not want human-like behavior or voices (ums, giggles, small talk); others wonder if a more human-feeling interface would help, but several push back: it’s a tool, not a friend.
Accuracy, Hallucinations, and Rigor
- Hallucinations and confident wrong answers are seen as the top technical problem, especially in math, theorem proving, and structured domains (databases, spreadsheets, law).
- Users want explicit “I don’t know” and clearer distinction between proven facts, conjecture, and guesses.
- Models sometimes double down on errors, invent tools or analyses they can’t do, and fail to signal uncertainty.
Memory, Context, and Long Conversations
- Complaints about short or opaque context windows, loss of earlier details, and degradation over long chats.
- Desire for: visible context usage, larger windows for Plus, better long-term memory that doesn’t duplicate or forget entries, and a way to “clear” or branch context without starting a new chat.
UX and Workflow Limitations
- Requests: better search/filter over past chats, forking threads, exporting entire conversations (e.g., markdown), editing text files directly, stable behavior across browsers, and improved mobile copy-paste.
- Users dislike verbosity, constant lists, overuse of em dashes and words like “comprehensive,” and frequent safety refusals or loops.
- Table formatting, image editing reliability, and canvas tools are seen as technically weak.
Bias, Alignment, and Training Opacity
- Concerns about opaque training data, hidden reasoning, and behavior shifts in cloud models.
- Observations that models favor longer, stronger language in arguments and tend to agree with the user rather than challenge nonsense.
Power-User and Niche Requests
- Style mimicry of the user’s own writing, math-competent models, fact-checking and bias metrics, raw vector access, Emacs integration, and using the same $20 subscription via API are all requested.