ai;dr

Use of LLMs in Writing vs “Slop”

  • Many distinguish between:
    • Low-effort: short prompt → long post → publish (seen as “slop”).
    • High-effort: long back-and-forth, heavy human editing and restructuring.
  • Some argue LLMs can sharpen thinking: questioning assumptions, finding gaps, steelmanning counterarguments.
  • Others doubt this, saying users feel smarter but rarely show concrete improvement.
  • A core objection: outsourcing thinking to a model, not just typing, is what people resent.

Effort, Trust, and the Broken Social Contract

  • A recurring theme: traditionally, writing takes more effort than reading; AI breaks that asymmetry.
  • Polished but generic prose is now a negative signal; typos, odd grammar, and “unpolished” style are becoming trust markers.
  • People report real frustration with LLM-fluffed corporate emails and docs: more words, less clarity.
  • However, some insist we should judge text by quality alone, not production method.

Detection Anxiety: Style Tells and Overreaction

  • Much discussion about “AI tells” (e.g., em dashes, certain paragraph cadences, “TED Talk” tone).
  • Some are altering their style (fewer em dashes, more rough edges) to avoid being misread as AI.
  • Others refuse to change, seeing that as ceding cultural ground to AI vendors.
  • General agreement that robust detection is hard and many self-proclaimed “LLM detectors” are overconfident.

Code, Docs, and Double Standards

  • Many happily use LLMs for code, tests, scaffolding, and documentation, claiming it’s “just for machines.”
  • Others push back: code and docs are also human communication; the same “effort” and “intention” arguments should apply.
  • Reports of AI-generated technical docs being confidently wrong deepen distrust and waste time.
  • Some leads see AI as enabling laziness and low-quality work (overlong design docs, noisy tickets, shallow research).

Information Economy and AI Mediation

  • Several expect that LLMs will become the main consumer of online text; humans will mostly see model summaries.
  • This incentivizes writing for the LLM (bland, factual, SEO-like), further homogenizing style.
  • Some propose reading only prompts (or author reputations) and ignoring AI-expanded prose.

Emotional and Cultural Loss

  • Multiple commenters describe AI as having “ruined” much online reading: voices now feel samey, parasocial writing less genuine.
  • Skepticism toward any polished writing increases cognitive load: readers constantly ask, “Did a person actually write this?”
  • There’s a desire for small, human-curated spaces and stronger norms around disclosure, without clear solutions.