Pixels to Torque: Figure AI's Helix 02 Solves Loco-Manipulation — The Hardest Problem in Home Robotics | Warmcore Tech
AI Architecture · Full-Body Autonomy Technical Deep Dive · April 8, 2026

Pixels to Torque: Figure AI's Helix 02 Solves Loco-Manipulation — The Hardest Unsolved Problem in Home Robotics

On January 27, 2026, Figure AI unveiled Helix 02: a single neural system that controls a humanoid's entire body — legs, torso, arms, and individual fingers — directly from camera pixels, with no human intervention. The robot completed a four-minute dishwasher task involving 61 continuous actions. Then it cleaned a living room. Then it jogged outdoors. The problem roboticists called "maybe impossible" has a working solution.

Neural network visualization — AI architecture for humanoid whole-body control loco-manipulation
Helix 02 replaces 109,504 lines of hand-engineered C++ code with a single 10-million-parameter neural network trained on 1,000+ hours of human motion data. Figure 03 hardware running Helix 02 has demonstrated 24/7 fully autonomous operation without human supervision. | Photo via Unsplash

01 — The Problem: Why Loco-Manipulation Defeated Robotics for Decades

There is a deceptively simple thing you do every morning without thinking about it. You walk from the bedroom to the kitchen, pick up a mug, fill it with water, carry it back, and set it on a table. Every step of that sequence requires something robotics researchers call loco-manipulation: the simultaneous, continuous coupling of locomotion and manipulation. Walk while carrying. Adjust balance while reaching. Recover from unexpected contact without stopping.

For decades, this has been one of the hardest unsolved problems in robotics — not because walking is hard, and not because object manipulation is hard, but because doing both together at the same time, through a body that has to balance continuously while its arms change the center of mass, resists the clean decomposition that engineers prefer. The standard workaround was a "stop-and-go" paradigm: walk to a destination, halt, stabilize, reach, grasp, then walk again. These handoffs are slow, brittle, and profoundly unlike how humans move. Any unexpected deviation — a wobble, a dropped object, an uneven floor — causes the whole sequence to fail.

"For decades, loco-manipulation has remained one of robotics' hardest unsolved problems. Not because either capability is hard alone, but because doing both together resists clean decomposition." — Figure AI, Helix 02 official announcement

Every humanoid robot demonstrated publicly before Helix 02 worked around this problem to some degree. Even impressive demonstrations — Boston Dynamics' backflips, Unitree's martial arts routines — are pre-planned motions with limited real-time feedback. They show what a robot can do when the environment cooperates exactly as expected. Helix 02 is the first system that addresses what happens when it doesn't.

61 Continuous actions in the 4-min dishwasher demo
109K Lines of C++ code replaced by one neural network
200Hz Real-time full-body joint control rate

02 — The Architecture: System 0, 1, and 2 — Three Speeds, One Body

The technical core of Helix 02 is a three-layer hierarchical architecture where each layer operates at the timescale appropriate to its function. The key insight — which Figure calls a "System 0, 1, 2" design — is that a robot body needs to reason at three very different speeds simultaneously: semantically slow, motorically fast, and physically instantaneous.

S2 7–9 Hz
System 2 — Semantic Reasoning Layer
Internet-pretrained vision-language model. Interprets scenes, understands natural language instructions, and plans high-level task sequences. Runs slowly because it can — "walk to the dishwasher and open it" doesn't need millisecond updates. Enables broad generalization: can reason about objects and instructions it has never seen before in training.
S1 200 Hz
System 1 — Whole-Body Visuomotor Policy
The "pixels-to-whole-body" layer. Takes visual input from head cameras, palm cameras, fingertip tactile sensors, and full-body proprioception — and outputs complete joint-level control of the entire robot simultaneously: legs, torso, head, arms, wrists, and individual fingers. Runs at 200 Hz to respond to the physical world in real time. This is the layer that was new in Helix 02 — the original Helix only controlled the upper body.
S0 1,000 Hz
System 0 — Physical Foundation Layer (New in Helix 02)
The critical new addition. A 10-million-parameter neural prior trained on 1,000+ hours of human motion data across 200,000+ parallel simulation environments. Replaces 109,504 lines of hand-engineered C++ with a single learned model of how human bodies move while maintaining balance and stability. Runs at 1 kHz because balance corrections need to be instantaneous. Figure describes it as "a foundation model for human-like whole-body control." Trained entirely in simulation, transfers directly to real hardware.

The three systems are trained end-to-end to communicate, not as independent modules with interfaces between them. S2 produces latent semantic representations that S1 reads as context for its visuomotor control. S1's outputs feed into S0's physical execution. The result is a robot that can "think slow" about what it's trying to do while "acting fast" to actually do it, and "balancing continuously" beneath both — without any of those three processes stopping for the others.

03 — The Demos: Dishwasher, Living Room, Outdoors

Figure has published three major Helix 02 demonstrations since January 2026, each expanding the scope of what the system can handle. Together they trace a clear trajectory from controlled kitchen environment to unstructured outdoor space.

January 27, 2026
The Kitchen: 4-Minute Dishwasher Cycle — 61 Actions, Zero Resets
Walk to dishwasher, open it, unload dishes, navigate across the room, stack items in cabinets, return, reload the dishwasher, start the cycle. Entirely autonomous, entirely onboard, no human intervention. Figure called it the longest-horizon, most complex task completed autonomously by a humanoid robot to date. The robot even used its foot to nudge the dishwasher door open when its hands were full.
March 2026
The Living Room: Spray, Wipe, Toss, Click
Walks into a messy living room. Picks up a spray bottle, wets a surface, wipes it with a towel. Collects toys and tosses them into a bin. Presses a TV remote. Repositions a pillow. All in a continuous shot — same neural system as the kitchen demo, no additional programming required. Helix 02 learns from data: adding task examples expands the repertoire without new code.
April 2026
Outdoors: Jogging at 2 m/s, Unsupervised 24/7 Operation
Figure 03 running Helix 02 demonstrated continuous unsupervised outdoor operation including jogging at approximately 2 meters per second. Separately, the company demonstrated 24/7 fully autonomous overnight runs without any human supervision — directly addressing the "what happens when no one is watching" question that all home robot deployments must answer.

04 — The Hardware: Figure 03's Sensor Suite Built for Helix 02

Helix 02 is not just a software breakthrough. The Figure 03 hardware was co-designed around the AI architecture's requirements — particularly the new sensor modalities that System 1's whole-body visuomotor policy needs to read from. Three additions stand out.

Fingertip tactile sensors sensitive to forces as low as three grams — roughly the weight of a paperclip — give the robot the ability to distinguish between a secure grip and an impending slip before the slip happens. This level of force feedback enables handling fragile objects that would be destroyed by a few hundred grams of excess force: glassware, eggs, pills, thin electronics. Palm-mounted cameras provide visual feedback when the robot's hands are occluded by the task itself — reaching into a bin, handling an object from below — situations where the head cameras lose useful information. 10 Gbps mmWave wireless data offload lets Figure 03 upload terabytes of operational data for continuous fleet-wide learning without requiring a wired connection.

Capability Helix (Original) Helix 02
Body coverage Upper body only Entire robot — legs, torso, arms, fingers
Foundation layer Not present System 0 at 1 kHz — human motion neural prior
Balance during manipulation Stop-and-stabilize required Continuous — walk while carrying, reach while moving
Tactile input Limited 3g-resolution fingertip sensors + palm cameras
Task horizon demonstrated Short sequences, single room 4-min full-kitchen cycle, multi-room, outdoor
New skill acquisition Requires new code or demonstrations Natural language — say what you want, robot does it
Overnight autonomous operation Not demonstrated Demonstrated — 24/7 without human supervision

05 — What Comes Next: 24/7 Operation, Factory Fleets, Home Pilots

Figure CEO Brett Adcock has outlined a specific 2026 roadmap built on the Helix 02 foundation. The targets are aggressive — but they are grounded in demonstrated capability rather than aspirational projections.

2026 — Now
Production line deployments: Figure 03 begins working in manufacturing environments, starting with operations comparable to or extending the BMW Spartanburg pilot where Figure 02 processed 90,000+ parts over 11 months.
2026 Q4
Home pilots for long-horizon tasks in completely unseen environments — the robot enters a house it has never been in before and performs complex household task sequences without prior facility mapping or task programming.
2027–2028
Robot-built robots: Figure's BotQ manufacturing facility, targeting 12,000 units annually scaling to 100,000 over four years, moves toward using Figure robots in its own production line — closing the loop between the robot and its own manufacture.
Ongoing
Hardware upgrades targeting superhuman speed and precision — Figure explicitly stated the goal is not just human-level capability but hardware that exceeds human performance on specific physical tasks in speed and accuracy.

06 — Why Helix 02 Matters Beyond Figure AI

The significance of Helix 02 is not limited to Figure AI's product roadmap. It is a proof of concept for an architectural approach — end-to-end neural control from sensors to actuators, without modular decomposition — that the entire robotics field is watching. If this approach generalizes, it changes the cost model for developing new robot behaviors from months of engineering work to days of data collection and training.

The traditional robotics paradigm required specialized engineers to program each new behavior: a PhD and hundreds of demonstrations per skill, as Figure's scaling curve chart illustrates. Helix 02 changes that equation. New capabilities are acquired by adding task examples to the training data — the model learns from them. Teach the robot to clean a kitchen; add living room examples; the bedroom follows without separate development. This is the same scaling dynamic that made large language models so powerful. Figure is applying it to physical intelligence.

For companies operating in the AI companion and high-fidelity interaction space, Helix 02 establishes a new reference point for what "full-body AI" means and what it makes possible. A robot that can walk into an unseen home and perform complex household tasks autonomously is not just an industrial tool. It is the physical substrate that companion AI has been waiting for: hardware capable of the kind of continuous, adaptive, whole-body presence that makes a robot feel like something more than a sophisticated appliance.

The home pilots planned for Q4 2026 will be the most important test. Factory environments are controlled. Homes are not. If Figure 03 running Helix 02 can perform reliably in the variable, unpredictable, human-occupied environments that consumer robots must handle — with fragile objects, irregular surfaces, unexpected obstacles, and people who don't behave like factory workers — the timeline to commercially viable home companion robots compresses significantly. That outcome is not guaranteed. But the architecture to attempt it exists now, publicly, in working hardware.

คุณอาจชอบเช่นกัน