A Year of Telepathy

Clinical results and technical questions

  • Commenters note the blog omits earlier reports that ~75–85% of threads in the first implant detached, raising questions about long‑term reliability and failure rates.
  • Some scrutinize usage graphs, seeing a one‑month spike then return to roughly linear growth; unclear if this reflects tech changes or just user behavior.
  • People ask how Neuralink handles scar tissue and glial response around electrodes, and whether there are genuine innovations here; no clear answers are surfaced in the thread.
  • Cursor control looks impressive but still “forced” to some viewers, with non‑direct paths and visible effort to avoid misclicks.

Comparison with prior BCI research

  • Multiple links show decades of similar brain–computer interface work (robot arms, sensory feedback, speech BCIs, gaming, non‑invasive control).
  • Several argue that proof‑of‑concept capabilities are not new; Neuralink’s differentiator is scale, PR, and willingness to push into human trials, not fundamental novelty.
  • Others stress that most labs avoid broader deployment because invasive BCIs remain immature, with unresolved longevity and scarring problems.

Risk, security, and autonomy trade-offs

  • A major thread contrasts abstract software/security concerns (hacking, hijacked implants, autonomy loss) with the lived reality of full paralysis.
  • Some say critics underweight the desperation of patients who might reasonably accept high risk for substantial autonomy.
  • Others insist software insecurity, hospital ransomware, and historical device failures make them unwilling to trust a brain‑controlling system, even if paralyzed.

Ethics: patients, access, and abandonment

  • There’s worry about “corporate cyborg parts”: implants becoming unsupported when companies pivot or die, leaving people stranded (bionic eye precedent cited).
  • Proposals include mandatory escrow of designs/source with public release if support ends, though many doubt current political and IP regimes would allow this.
  • Access and cost loom large: disabled people already struggle with basic needs; some fear tools that are life‑changing but only for the wealthy few.

Animal testing debate

  • Reports of roughly 1,500 animals killed spark intense argument.
  • One camp says choosing between animals and restoring autonomy to humans is ethically straightforward.
  • Another distinguishes high‑quality, carefully designed animal studies from sloppy or wasteful experiments, arguing that sheer numbers don’t excuse poor practice.
  • Some point out the inconsistency of meat‑eaters condemning animal testing; vegetarians/vegans in the thread still judge Neuralink’s record “horrific.”

Trust, regulation, and Musk’s role

  • Many express deep distrust tied to Musk: perceived hostility to regulation, past misleading demos, PR‑driven timelines, and broader political actions (especially dismantling oversight agencies and foreign aid programs).
  • Others counter that major technological advances (EVs, reusable rockets) often came from similarly “unhinged” founders and argue that Neuralink should be judged on outcomes, not personality.
  • A skeptical subgroup suggests disabled patients are doubling as “cobayes and free marketing,” given Neuralink’s longer‑term ambitions for mass BCIs, cyborg enhancement, or human–AI competition.

Language, branding, and sci‑fi fears

  • The product name “Telepathy” and talk of “mind control” are criticized as sensational marketing, analogous to calling LLMs “AI” rather than “ML.”
  • Several commenters invoke sci‑fi (Greg Egan, Banks, Black Mirror) to explore possibilities of hacked or coercive implants, torture via neural laces, or overwritten agency.
  • There is specific concern about future LLM‑mediated decoding “speaking” for patients in ways that might diverge from their actual intentions.

Meta: tone of the discussion

  • Some lament that discourse is dominated by US politics and Musk hatred rather than the technology or patients’ experiences.
  • Others respond that with an invasive medical device, the founder’s ethics, regulatory stance, and political power are intrinsically part of the risk calculus.