I put my heart and soul into this AI but nobody cares

AI, “Mind Control,” and Vulnerability

  • Several comments frame AI images/videos as the latest “mind-control” tech, designed to manipulate emotions at scale and serve powerful economic/political actors.
  • Debate over who is vulnerable: some point to the poor and socially isolated; others say the truly “immune” are those who control distribution platforms.
  • A few argue that maintaining real-life relationships and avoiding public social feeds is the only realistic defense.

Social Media Slop and Bot Farms

  • Many describe Facebook as saturated with AI-generated “engagement farms”: low-effort emotional bait with fake profiles, generic photos, and shallow comments.
  • Some say this is indistinguishable from most normal user behavior, and note that platforms profit via inflated engagement metrics and ad revenue.
  • Others think this is just an extension of old clickbait; AI only cheapens and scales it.

Authenticity, Detection, and the Article Itself

  • Multiple commenters say they initially assumed the article was AI-written due to repetitive descriptions and “flat” style.
  • When the author appears to insist it’s human-written, some remain skeptical; others attribute the style to accessibility practices (describing images for screen readers).
  • There is interest but little faith in AI detectors; reported false positives on ordinary comments reinforce the view that current tools are unreliable.
  • Several argue that whether the content is AI- or human-generated matters less than its manipulative power.

Critical Thinking, Education, and Religion

  • One camp insists critical thinking should be central in schools to help people resist scams and misinformation.
  • Others counter that emotion trumps reason in practice, so rational skills alone won’t protect people.
  • The thread branches into a long argument over whether critical thinking reduces religiosity, whether religion fulfills psychological/social needs, and whether a nihilistic worldview is livable—no consensus emerges.

Political and Societal Impact

  • Example from India: a deepfake of a major politician allegedly caused a significant vote shift before being debunked.
  • Private, encrypted platforms (e.g., WhatsApp groups) are portrayed as powerful rumor mills with limited oversight, sometimes linked to real-world violence.
  • Several worry that constant exposure to fake sob stories “farms empathy,” humiliating well-meaning people and eventually numbing genuine compassion.

Platform Economics and Algorithms

  • Commenters link AI slop to get-rich-quick ecosystems: YouTube “how to make money with AI” gurus, cheap labor in low-income countries, and microtransaction systems (stars, gifts, bits).
  • Some note Facebook’s role in subsidizing data access in poorer regions, effectively farming attention at scale and creating fertile ground for scams and AI spam.
  • Many blame engagement-optimized recommendation algorithms more than AI itself: whatever drives clicks—rage, pity, or awe—gets amplified.

User Responses and Coping Strategies

  • Some adopt blanket cynicism: treat everything online as fake and disengage emotionally.
  • Others advocate simply abandoning platforms that shovel “useless shit,” but acknowledge most casual users won’t.
  • There’s concern that as people respond by distrusting everything, society’s ability to share facts, sustain empathy, and act collectively may erode further.