These AI robots can now move even faster than humans as figure AI just released the second demo of its newest Helix AI system packed with four new tech breakthroughs plus a special bonus but how close to AGI are they now and what can they do as a result to answer that figure tested helix's system one visual motor control in several real world scenarios demoing the following four new tech breakthroughs number one implicit stereo Vision the first important intelligence feature of the helix S1 is its adoption of implicit stereo Vision giving the system a rich three-dimensional
understanding of its environment and unlike its predecessor which relied on monocular visual input the upgraded S1 uses a stereovision backbone paired with a multiscale feature extraction Network this architectural shift allows the robot to merge visual data from two cameras into a cohesive depth aware picture before processing it through a cross attention Transformer the result is the robots except ability to perceive fine environmental details while also maintaining a broader understanding of the scene such as the layout of a bustling conveyor belt in a factory to prove this the robot's stereo Vision capability was put to the
test in figur logistics demo with packages of varying sizes shapes and weights streaming across a conveyor belt requiring precise depth perception to grasp and reorient them correctly and the data from these tests showed a 60% increase in throughput compared to non-stereo baselines with the system even generalizing to Flat envelopes that it wasn't explicitly trained on and this humanlike spatial awareness serves as a key visual Foundation to enable the next breakthrough that Helix has to offer number two multiscale visual representation this second major advancement Builds on top of helix's stereo Vision Foundation allowing Helix S1 to
capture both granular details and highlevel contextual cues simultaneously plus rather than processing images from each camera independently the system fuses stereo inputs into a multiscale stereo Network producing a compact set of visual tokens that feed into the Transformer without increasing its overall computational load this balance of detail and context enables more reliable control whether the robot is picking up a tiny parcel or navigating a crowded workspace and in the logistics use case this capability is invaluable because packages on a conveyor belt often vary wildly with some items being small and rigid While others might be
large and deformable forcing the robot to decide the optimal Gras point and Method on the fly but the new multiscale approach ensures Helix doesn't lose sight of the bigger picture while zeroing in on critical specifics like a shipping labels orientation and the results show that the addition of multiscale feature extraction significantly boosts the effective throughput referred to as TF which is a metric that compares the robot's package handl speed to human package handling speed all of this enables the Next Great Leap in robot intelligence number three learned visual proprioception this third breakthrough tackles another long-standing
challenge in robotics which is successfully scaling a single policy across multiple humanoid robots this is partially because Hardware variations like slight differences in sensor calibration or joint responses can disrupt performance when a policy trained on one robot is applied to another and while manual calibration used to be the traditional fix it just doesn't scale for an entire fleet but Helix S1 sidesteps this with a self-calibration system that estimates the six-dimensional poses of its end defectors such as its hands using only onboard visual input with no external tools required as a result this feature allows figure
to deploy the same policy trained on a single robot across multiple units with minimal performance drop and despite small Hardware works the robots demonstrate consistent dexterity flipping packages to expose labels and transferring them between conveyors at high speed importantly this cross robot transfer capability slashes downtime and calibration costs to make large scale deployments feasible but what comes next pushes Helix into superhuman territory number four sport mode this fourth break through catapults Helix S1 into faster than human territory by using a simple yet ingenious test time technique where figure speeds up the robot's actions by resampling
its output action chunks which are sequences of movements generated at 200 htz for example a chunk representing a 1second trajectory can be compressed to 0.8 seconds and executed at the original rate yielding a 20% speed boost without retraining the model in testing figer pushed this to a 50% speed up achieving a TF value greater than one meaning the robot outpaced its human demonstrators and sport mode enables helix's AI to handle packages with Incredible efficiency maintaining High success rates even as throughput sword however the demo also showed that while a 50% Speed Up maximizes Performance pushing
Beyond this upper threshold sacrifices precision and requires frequent resets but the demo still highlights a tantalizing Prospect humanoid robots that don't just match human speed but surpass it all while retaining the dexterity needed for complex tasks like label orientation and here's why it all matters so much these advancements stereo Vision multiscale representation visual proprioception and sport mode aren't just incremental upgrades but a paradigm shift by solving core challenges like depth perception contextual awareness Fleet scalability and speed figure has unlocked a future where humanoid robots can seamlessly integrate into human workflows the logistics demo is just
the beginning these generic improvements to Helix S1 will enhance every use case figure pursues from manufacturing to healthc care and the numbers back up the hype a TF above 1.1 means Helix is already 10% faster than its human trainers in some scenarios with sport mode pushing that edge further what's more is that the stereo model's 60% throughput jump coupled with the proprioception module's cross robot consistency both signal that this technology is ready for prime time as for price figure ai's CEO suggests their AI robots are already generating Revenue possibly at a premium for early adopters
like BMW but for broader markets such as Logistics or home use economies of scale could eventually push the price closer to $50,000 as production ramps up aligning with adcock's vision of deploying 100,000 robots in the next four years but another price strategy may be a subscription-based model which could make humanoid robots even more accessible this could allow businesses and consumers to just pay a monthly fee in a leasing structure instead of a purchasing model along this trajectory figure might follow the robotics as a service model where they'd essentially rent out robots for tasks like Warehouse
sorting or package handling with pricing based on hours of use or tasks completed and Beyond this figure would also potentially maintain ownership of the data which they could monetize and then pass down savings to users lowering the entry barrier for smaller firms and consumers while ensuring more predictable revenue for figure and if BMW and other Enterprise clients are already seeing value figure could even introduce tiered pricing offering premium AI capabilities faster updates or dedicated support at higher subscription levels anyways tell us how much you'd be willing to pay for one of these robots in the
comments below and click here to listen to figures CEO talk about his plans for 2025 and Beyond