The case against conversational interfaces
Screens, Childhood, and “Computerized Humans”
- Several commenters reflect on how rapidly humans adapted to screens and remotes, arguing that people clearly like screen-based, finger-driven control and won’t abandon it for voice.
- Debate over early childhood exposure: Waldorf-style “no tech before ~6” is praised for supporting imagination and easier “switching off,” but others say early screen use (especially games) was life‑saving, opening access to ideas beyond hostile or anti-intellectual environments.
- Consensus that “screens aren’t the problem, bad content and absent parenting are”; quality and boundedness of games/TV (e.g., Ghibli vs. Cocomelon, Factorio vs. gacha games) matter more than the medium.
Voice / Conversational Interfaces: Where They Work
- Seen as a good secondary channel: setting timers, controlling lights, checking weather, simple reminders, querying smart speakers while hands/eyes are busy.
- Useful for rarely used or complex functions where users know what they want but not how to do it (BI queries, “book a flight around 7pm Friday within this budget,” ad‑hoc scripting, advanced app features).
- Works well as a proxy for a human assistant: small businesses or executives issuing high‑level intents instead of clicking through complex tools.
…and Where They Fail
- Repeated claims that speaking is slower, more tiring, and socially impractical; catastrophic in shared or noisy environments and for dense, precise tasks (coding, driving, gaming, flight search, form-filling).
- Serial, low-bandwidth, and memory-heavy: worse than visually scanning options, comparing many alternatives, or making fine-grained adjustments (car controls, shopping, flight selection).
- Real-world chat/“Copilot Studio” UIs often degrade UX: linear, confirmation-heavy flows that replace a simple form/datepicker with slow back-and-forth.
Augmentation, Not Replacement
- Strong alignment with the article’s core view: natural language should augment, not replace, GUIs and keyboard/mouse.
- Many want an OS-level agent that observes context and supports “telepathic” automation of routine tasks, but without constantly guessing and rearranging interfaces.
- Pushback against “tools trying to be smart”: predictive UIs, autocorrect, algorithmic feeds, and hidden options are seen as hostile to user agency; stable spatial interfaces and explicit commands are preferred.
LLMs, Ambiguity, and Hybrid Design
- Some praise LLMs as mediators that can turn vague human intents into exact commands and summarize complex outputs; others stress nondeterminism, hallucinations, and security risks.
- Natural language is framed by some as a poor “data-transfer mechanism” but by others as a powerful way to negotiate intent, much like working with senior engineers or travel agents.
- Broad agreement that the future is multimodal: keyboard shortcuts, pointing, and structured UIs for speed and precision, with conversational layers for discovery, delegation, and edge cases.