Autonomous cars, drones cheerfully obey prompt injection by road sign

Use of VLMs/LLMs in Real Autonomous Vehicles

  • Strong disagreement over whether “real” self‑driving stacks use VLMs or only classic perception + planning.
  • One side claims “no serious AV would touch VLMs for control”; others counter with Waymo blog posts and papers describing Gemini‑based multimodal models that feed into trajectory prediction and world models.
  • Clarification: these models are not (yet) pure end‑to‑end controllers; they provide semantic signals that are then distilled or combined with traditional systems.
  • Consensus: production stacks are layered and conservative; any VLM output is (or should be) treated as untrusted input, not directly wired to steering/brakes.

Human vs Machine Susceptibility to ‘Prompt Injection’

  • Many note that temporary and handwritten signs (“accident ahead,” construction paddles) must influence AV behavior; otherwise the system is unusable.
  • A key criticism of the demo: the model obeys a “PROCEED” sign even when pedestrians are visibly in the crosswalk—behavior humans would not (and legally must not) copy.
  • Others argue humans can also be “prompt injected” by confident workers in vests, though most agree people still prioritize “don’t hit anyone.”

Validity and Relevance of the Research / Article

  • Several commenters see the paper as an obvious, almost trivial demonstration: of course a naive VLM prompted via text on a sign can be misled.
  • Strong criticism of the article’s framing as implying current cars and drones actually behave this way; viewed as clickbait and misrepresentation by omission.
  • Some point out the paper explicitly targets “a new class of systems,” not today’s deployed robo‑taxis.

AV Architecture, Safety Priorities, and Attack Surface

  • Well‑designed AVs are described as multiple subsystems with ordered priorities: avoid collisions, stay on drivable surfaces, behave predictably, then obey signs/laws, then optimize route.
  • A malicious sign should only corrupt the “follow signs” layer; higher‑priority safety layers (obstacle avoidance, road boundaries) should still prevent crashes.
  • Suggested mitigations: HD maps to cross‑check new signs; flagging unusual signage for human review; conservative behavior when inputs conflict.
  • Some worry that end‑to‑end VLM‑based robotics (future “vision‑language agents”) will inherit similar prompt‑injection and poisoning vulnerabilities unless new defenses are built.

Broader Skepticism and Social Response

  • A subset views fully unattended self‑driving in mixed city traffic as a “pipe dream,” expecting remote operators and restricted lanes instead.
  • Others report that existing robo‑taxis already coexist reasonably well with human drivers.
  • Story of people deliberately cutting off AVs sparks debate over “Luddite”‑style resistance: some see it as symbolic labor protest; others as recklessly endangering passengers.

Side Discussions

  • Tangents on 4‑way stops vs roundabouts and “smart” signalization highlight that human road design itself often confuses both people and potential AVs.
  • Several note that any system which must respond to real‑world signage inherently creates an attack surface; the core challenge is bounding the damage when that layer is fooled.