Helix: A vision-language-action model for generalist humanoid control
Model architecture & control approach
- Commenters highlight the two-model design: a slower 7B vision‑language model (7–9 Hz) producing latent “intent” vectors, and a small 80M visuomotor model (200 Hz) mapping those to joint actions.
- Several note this mirrors established robotics practice: high‑level planning plus low‑level controllers, with traditional controllers/motor drivers still handling torque, balance, and PWM-level signals.
- Some wonder how the latent interface is structured (custom “control tokens,” coordinates, learned codebook, etc.) and how the small model fuses its own sensory state with that latent guidance.
Training, generalization & “first time seeing objects”
- Debate over what “first time you’ve seen these objects” means: unseen in robot videos vs unseen visually at all vs known from internet pretraining.
- Some assume standard train/validation split; others liken it to a child recognizing an apple from prior textual description.
- A few are skeptical of the breadth of the claimed zero‑shot generalization, given training appears focused on household tasks.
Demo authenticity, environment & limitations
- Multiple people distrust polished robotics demos generally: questions about staging, retries, and whether parts are pre‑scripted or sped up.
- The minimalist, sterile kitchen is seen as both visually slick and much easier than real clutter; several request tests in uncontrolled spaces (homes, construction sites, warehouses).
- Others note that capabilities like chopping onions, dealing with loose skins, or complex in‑hand manipulation are conspicuously absent.
Household robots, AR assistants & daily life
- Strong interest in domestic automation (laundry folding, cleaning, cooking), but split views on desirability: some see it as liberation, others as giving up meaningful care of home/family.
- Cost comparisons are made to human cleaners and existing laundry services; some doubt humanoids will be cost‑competitive soon.
- Several propose near‑term “AI as brain, human as hands” systems: AR guidance for groceries, repairs, recipes, and home organization, versus full physical autonomy.
Human interaction, aesthetics & anthropomorphism
- The robots’ black, faceless, “Bond villain intern” look is widely called sinister and uncanny, especially in a domestic setting.
- Repeated “eye contact” after handoffs is perceived as forced anthropomorphism to impress investors, though some argue such cues matter for human‑robot interaction.
- A few find the lack of speech dehumanizing; others say talking and gaze would be useful social signals.
Safety & reliability
- Concern over physical safety: motor torque, inertia, falling robots, and unsafe behaviors in kitchens or around children/pets.
- Suggestions include torque/velocity limits, independent safety controllers, and force‑sensing co‑robots, but others argue that for open‑ended humanoids the hazard space is too large for traditional SIL‑style safety engineering.
- Some note today’s slow, cautious movements reduce immediate risk but do not solve behavioral safety.
Warfare & misuse
- Several connect this directly to lethal autonomy: robots manning howitzers, “stabby” drones, ethnic targeting, and swarms as new WMDs.
- Others point out that autonomous and semi‑autonomous killing systems already exist (drones, AI target selection), questioning what “conversation” is left to start.
Technical open questions
- Questions about hand design: degrees of actuation, compliance, and ultimate in‑hand dexterity (e.g., Rubik’s cube–level tasks).
- Curiosity about how 3D space is represented: explicit depth sensors vs learned depth from RGB, and how coordination between multiple robots is implemented.
- Some note the 200 Hz control rate is high but plausible for low‑level control; others ask whether a single unified multimodal model could eventually replace the two‑tier architecture.
Business model, infrastructure & skepticism
- People ask why, if it can “pick up anything,” it isn’t already deployed at scale in industrial picking (e.g., Amazon), and suggest demos are aimed at boosting valuation.
- There’s debate over whether models and inference are truly “on‑robot”: the marketing site implies significant cloud/offboard compute.
- Ancillary gripe: their self‑hosted video streaming performs poorly; several prefer YouTube/real CDNs.