Xkcd 1425 (Tasks) turns ten years old today

Bird vs. park: which task is actually harder now?

  • Many note the comic’s reversal: bird detection is now “trivial” with off‑the‑shelf CNN/YOLO models; a few lines of code can detect birds in images.
  • Others point out constraints the comic ignored: doing this on a 2014 phone, or without today’s models and tooling, would still have been hard.
  • The “is this in a national park?” task hides complexity: GPS inaccuracy, bad reception, boundary ambiguity (rivers, cities in parks), changing geography, and unclear requirements for “in” vs “near” a park.
  • GIS lookup itself (point‑in‑polygon against public park polygons) is technically straightforward now, but still depends on decades of prior work and datasets.

Progress in ML and shifting goalposts

  • Commenters reflect on how impressive GANs and early convnets looked a decade ago versus today’s standards.
  • Some argue expectations for AGI and the Turing test have been continuously “moved”; others say tests must naturally get stricter as systems improve.
  • There’s debate whether current LLMs count as anything close to AGI, with some insisting core capabilities (strong reasoning, robust math, one‑shot learning) are still missing.

Turing test, “real thinking,” and philosophy

  • Long sub‑thread on whether LLMs “really think” or just simulate language; references to Turing’s original framing, functionalism, and the Chinese Room.
  • Some stress that questions like “does it think?” are partly semantic; others see deep qualitative gaps between human and model reasoning, despite superficial fluency.

Jobs, productivity, and “good enough” automation

  • Disagreement on how many current jobs could be replaced today: some see many roles as trivially automatable; others say almost no jobs are fully replaceable yet.
  • Concern that AI will be used where it’s cheaper but worse (translation, customer service), leading to degraded quality but higher profits.
  • Others see AI enabling new jobs in places that previously couldn’t afford human labor at all.

Why similar‑looking software tasks differ by orders of magnitude

  • Many emphasize the comic’s core point: non‑technical stakeholders often can’t tell which tasks are “move a chair” versus “move the toilet and plumbing.”
  • House and plumbing analogies are used to explain why tiny‑sounding requirements (e.g., “robust MI,” “small UI tweak,” second optimistic‑update path) can imply major re‑architecture.

LLMs’ capabilities and sharp edges

  • LLMs excel at data cleanup, metadata normalization, simple CV (vision APIs), and rapid prototyping; they’re described as “ultimate interns.”
  • Failure modes are highlighted: hallucinations, weak arithmetic, reluctance to say “I don’t know,” and trouble with tokenization‑sensitive tasks (spelling “STRAWBERRY,” rendering “HELLO” in vegetables, negation like “no cheese”).
  • Some see these as inherent to current token‑based architectures; others think better training, tools, or hybrid systems can mitigate many issues.

Infrastructure, maturity, and “easy vs hard”

  • Multiple comments stress that both tasks (GPS/GIS and bird detection) became “easy” only after huge investments in satellites, GIS standards, datasets, GPUs, and ML research.
  • The comic is read less as a statement about vision per se and more as a reminder that perceived vs actual difficulty in software depends heavily on invisible infrastructure and historical context.