| When large language models (LLMs) can already write copy and answer questions, another, deeper technological revolution is quietly unfolding—teaching machines to truly touch and interact with the physical world. NVIDIA executive Spencer Huang, son of Jensen Huang, recently shared his vision of the robotics revolution. Why does he call it the next trillion-dollar industry? How can robots learn delicate operations—like picking up an egg without breaking it? And how far are we from having robots in our homes? At the AI Reinvent private session, Huang broke down the toughest challenges in Physical AI, while sketching a roadmap from industrial robotic arms to humanoid home assistants. “People say robotics may become the next trillion-dollar heavy industry. Personally, I think that figure is still conservative,” said Spencer Huang. 01. What Robots Need to Learn Are Things Humans Do Instinctively The most fascinating aspect of general-purpose robots is that they seem able to manipulate the world like humans. But, as Huang pointed out, this task is ten times harder than we imagine. Humans naturally know how to grab a cup by the rim or fold clothes by aligning corners. But such “common sense” has never been written into manuals. For a robot, handling an egg is not about hearing “be gentle,” but about discovering the precise force and grip point that prevent it from breaking. The bottleneck, Huang stressed, lies in the data gap of Physical AI. While LLMs learn from centuries of human writing, there are no large-scale “instruction books” on how to physically interact with the world. 02. Why Simulation Is the Lifeline “You can’t expect a miner to teach a robot to dig coal, or a chef to train it to toss a wok. Real-world data is scarce and costly,” Huang explained. That’s where simulation comes in—as a virtual training ground for robots. Synthetic Data at Scale: In simulators, a robot can pick up cups a million times, with shape, weight, and surface texture changing each round. Through domain randomization, robots “see the world” in endless variations. Safe, Cheap Trial and Error: In reality, a broken part may cost thousands of dollars. In simulation, even if a robot smashes a virtual table, a reset takes seconds. Tools like NVIDIA Omniverse enable robots to learn by failing safely until they master the right amount of force. Huang emphasized the importance of physics fidelity: if simulated friction and gravity don’t match reality, robots that excel virtually will stumble in the real world. 03. How Robots Get Smarter: Two Learning Paths The NVIDIA team is advancing two major approaches: Reinforcement Learning: Like children learning to ride bikes, robots improve through trial and error—falling, adjusting, and stabilizing. With sensors, they can now adjust grip strength in real time, making it “smart trial and error.” Behavior Cloning: Robots mimic humans directly. Wearing VR headsets and gloves, people demonstrate tasks, which robots then replicate. Huang envisions a future where robots will learn from watching cooking videos—moving from imitation to reasoning and generalization. 04. World Foundation Models: Giving Robots “Imagination” “Imitation alone isn’t enough. Robots must learn to create,” Huang noted. World foundation models are now being developed to inject imagination into robotics: Augmenting Realism: Generating kitchens with cracked cups and spilled water, so robots adapt to imperfection. Inventing New Tasks: From a single instruction—“put the apple on the plate”—robots can generate new motions and apply them flexibly. Cross-Domain Transfer: Testing if a robot trained in kitchens can handle factories, much like an exam in a new classroom. 05. The Next Decade: From Tools to Assistants Huang compared robot evolution to a spectrum: Now: Specialized robots dominate—factory welders, farm harvesters, surgical assistants—masters of one task with near-perfect precision. Next 10 Years: General-purpose robots will emerge—like general practitioners, capable of multiple household tasks, from washing dishes to assisting the elderly. Timeline: In 3–5 years, household robots may pick up socks or wipe tables. In 10 years, bimanual robots could assemble furniture—if breakthroughs in tactile sensing and scalable data are achieved. 06. Will Robots Replace Humans? “More likely, it’s collaboration,” Huang said. Humans excel at creativity and decision-making, while robots thrive in repetitive labor. A factory worker, for instance, may no longer bend over to move parts, but instead supervise and guide robots: “Be gentler here.” In Huang’s view, robots won’t replace us but will liberate us—much like calculators didn’t make humans less intelligent, but allowed us to solve more complex problems. “This transformation will be even deeper than the computer revolution,” he concluded. “Machines that can truly touch the world will redefine how we live and interact.” |


