For a generation of scientists raised watching Star Wars, theres a disappointing lack of C-3PO-like droids wandering around our cities and homes. Where are the humanoid robots fuelled with common sense that can help around the house and workplace?
Rapid advances in artificial intelligence (AI) might be set to fill that hole. I wouldnt be surprised if we are the last generation for which those sci-fi scenes are not a reality, says Alexander Khazatsky, a machine-learning and robotics researcher at Stanford University in California.
From OpenAI to Google DeepMind, almost every big technology firm with AI expertise is now working on bringing the versatile learning algorithms that power chatbots, known as foundation models, to robotics. The idea is to imbue robots with common-sense knowledge, letting them tackle a wide range of tasks. Many researchers think that robots could become really good, really fast. We believe we are at the point of a step change in robotics, says Gerard Andrews, a marketing manager focused on robotics at technology company Nvidia in Santa Clara, California, which in March launched a general-purpose AI model designed for humanoid robots.
At the same time, robots could help to improve AI. Many researchers hope that bringing an embodied experience to AI training could take them closer to the dream of artificial general intelligence AI that has human-like cognitive abilities across any task. The last step to true intelligence has to be physical intelligence, says Akshara Rai, an AI researcher at Meta in Menlo Park, California.
But although many researchers are excited about the latest injection of AI into robotics, they also caution that some of the more impressive demonstrations are just that demonstrations, often by companies that are eager to generate buzz. It can be a long road from demonstration to deployment, says Rodney Brooks, a roboticist at the Massachusetts Institute of Technology in Cambridge, whose company iRobot invented the Roomba autonomous vacuum cleaner.
There are plenty of hurdles on this road, including scraping together enough of the right data for robots to learn from, dealing with temperamental hardware and tackling concerns about safety. Foundation models for robotics should be explored, says Harold Soh, a specialist in humanrobot interactions at the National University of Singapore. But he is sceptical, he says, that this strategy will lead to the revolution in robotics that some researchers predict.
The term robot covers a wide range of automated devices, from the robotic arms widely used in manufacturing, to self-driving cars and drones used in warfare and rescue missions. Most incorporate some sort of AI to recognize objects, for example. But they are also programmed to carry out specific tasks, work in particular environments or rely on some level of human supervision, says Joyce Sidopoulos, co-founder of MassRobotics, an innovation hub for robotics companies in Boston, Massachusetts. Even Atlas a robot made by Boston Dynamics, a robotics company in Waltham, Massachusetts, which famously showed off its parkour skills in 2018 works by carefully mapping its environment and choosing the best actions to execute from a library of built-in templates.
For most AI researchers branching into robotics, the goal is to create something much more autonomous and adaptable across a wider range of circumstances. This might start with robot arms that can pick and place any factory product, but evolve into humanoid robots that provide company and support for older people, for example. There are so many applications, says Sidopoulos.
The human form is complicated and not always optimized for specific physical tasks, but it has the huge benefit of being perfectly suited to the world that people have built. A human-shaped robot would be able to physically interact with the world in much the same way that a person does.
However, controlling any robot let alone a human-shaped one is incredibly hard. Apparently simple tasks, such as opening a door, are actually hugely complex, requiring a robot to understand how different door mechanisms work, how much force to apply to a handle and how to maintain balance while doing so. The real world is extremely varied and constantly changing.
The approach now gathering steam is to control a robot using the same type of AI foundation models that power image generators and chatbots such as ChatGPT. These models use brain-inspired neural networks to learn from huge swathes of generic data. They build associations between elements of their training data and, when asked for an output, tap these connections to generate appropriate words or images, often with uncannily good results.
Likewise, a robot foundation model is trained on text and images from the Internet, providing it with information about the nature of various objects and their contexts. It also learns from examples of robotic operations. It can be trained, for example, on videos of robot trial and error, or videos of robots that are being remotely operated by humans, alongside the instructions that pair with those actions. A trained robot foundation model can then observe a scenario and use its learnt associations to predict what action will lead to the best outcome.
Google DeepMind has built one of the most advanced robotic foundation models, known as Robotic Transformer 2 (RT-2), that can operate a mobile robot arm built by its sister company Everyday Robots in Mountain View, California. Like other robotic foundation models, it was trained on both the Internet and videos of robotic operation. Thanks to the online training, RT-2 can follow instructions even when those commands go beyond what the robot has seen another robot do before1. For example, it can move a drink can onto a picture of Taylor Swift when asked to do so even though Swifts image was not in any of the 130,000 demonstrations that RT-2 had been trained on.
In other words, knowledge gleaned from Internet trawling (such as what the singer Taylor Swift looks like) is being carried over into the robots actions. A lot of Internet concepts just transfer, says Keerthana Gopalakrishnan, an AI and robotics researcher at Google DeepMind in San Francisco, California. This radically reduces the amount of physical data that a robot needs to have absorbed to cope in different situations, she says.
But to fully understand the basics of movements and their consequences, robots still need to learn from lots of physical data. And therein lies a problem.
Although chatbots are being trained on billions of words from the Internet, there is no equivalently large data set for robotic activity. This lack of data has left robotics in the dust, says Khazatsky.
Pooling data is one way around this. Khazatsky and his colleagues have created DROID2, an open-source data set that brings together around 350 hours of video data from one type of robot arm (the Franka Panda 7DoF robot arm, built by Franka Robotics in Munich, Germany), as it was being remotely operated by people in 18 laboratories around the world. The robot-eye-view camera has recorded visual data in hundreds of environments, including bathrooms, laundry rooms, bedrooms and kitchens. This diversity helps robots to perform well on tasks with previously unencountered elements, says Khazatsky.
When prompted to pick up extinct animal, Googles RT-2 model selects the dinosaur figurine from a crowded table.Credit: Google DeepMind
Gopalakrishnan is part of a collaboration of more than a dozen academic labs that is also bringing together robotic data, in its case from a diversity of robot forms, from single arms to quadrupeds. The collaborators theory is that learning about the physical world in one robot body should help an AI to operate another in the same way that learning in English can help a language model to generate Chinese, because the underlying concepts about the world that the words describe are the same. This seems to work. The collaborations resulting foundation model, called RT-X, which was released in October 20233, performed better on real-world tasks than did models the researchers trained on one robot architecture.
Many researchers say that having this kind of diversity is essential. We believe that a true robotics foundation model should not be tied to only one embodiment, says Peter Chen, an AI researcher and co-founder of Covariant, an AI firm in Emeryville, California.
Covariant is also working hard on scaling up robot data. The company, which was set up in part by former OpenAI researchers, began collecting data in 2018 from 30 variations of robot arms in warehouses across the world, which all run using Covariant software. Covariants Robotics Foundation Model 1 (RFM-1) goes beyond collecting video data to encompass sensor readings, such as how much weight was lifted or force applied. This kind of data should help a robot to perform tasks such as manipulating a squishy object, says Gopalakrishnan in theory, helping a robot to know, for example, how not to bruise a banana.
Covariant has built up a proprietary database that includes hundreds of billions of tokens units of real-world robotic information which Chen says is roughly on a par with the scale of data that trained GPT-3, the 2020 version of OpenAI's large language model. We have way more real-world data than other people, because thats what we have been focused on, Chen says. RFM-1 is poised to roll out soon, says Chen, and should allow operators of robots running Covariants software to type or speak general instructions, such as pick up apples from the bin.
Another way to access large databases of movement is to focus on a humanoid robot form so that an AI can learn by watching videos of people of which there are billions online. Nvidias Project GR00T foundation model, for example, is ingesting videos of people performing tasks, says Andrews. Although copying humans has huge potential for boosting robot skills, doing so well is hard, says Gopalakrishnan. For example, robot videos generally come with data about context and commands the same isnt true for human videos, she says.
A final and promising way to find limitless supplies of physical data, researchers say, is through simulation. Many roboticists are working on building 3D virtual-reality environments, the physics of which mimic the real world, and then wiring those up to a robotic brain for training. Simulators can churn out huge quantities of data and allow humans and robots to interact virtually, without risk, in rare or dangerous situations, all without wearing out the mechanics. If you had to get a farm of robotic hands and exercise them until they achieve [a high] level of dexterity, you will blow the motors, says Nvidias Andrews.
But making a good simulator is a difficult task. Simulators have good physics, but not perfect physics, and making diverse simulated environments is almost as hard as just collecting diverse data, says Khazatsky.
Meta and Nvidia are both betting big on simulation to scale up robot data, and have built sophisticated simulated worlds: Habitat from Meta and Isaac Sim from Nvidia. In them, robots gain the equivalent of years of experience in a few hours, and, in trials, they then successfully apply what they have learnt to situations they have never encountered in the real world. Simulation is an extremely powerful but underrated tool in robotics, and I am excited to see it gaining momentum, says Rai.
Many researchers are optimistic that foundation models will help to create general-purpose robots that can replace human labour. In February, Figure, a robotics company in Sunnyvale, California, raised US$675 million in investment for its plan to use language and vision models developed by OpenAI in its general-purpose humanoid robot. A demonstration video shows a robot giving a person an apple in response to a general request for something to eat. The video on X (the platform formerly known as Twitter) has racked up 4.8 million views.
Exactly how this robots foundation model has been trained, along with any details about its performance across various settings, is unclear (neither OpenAI nor Figure responded to Natures requests for an interview). Such demos should be taken with a pinch of salt, says Soh. The environment in the video is conspicuously sparse, he says. Adding a more complex environment could potentially confuse the robot in the same way that such environments have fooled self-driving cars. Roboticists are very sceptical of robot videos for good reason, because we make them and we know that out of 100 shots, theres usually only one that works, Soh says.
As the AI research community forges ahead with robotic brains, many of those who actually build robots caution that the hardware also presents a challenge: robots are complicated and break a lot. Hardware has been advancing, Chen says, but a lot of people looking at the promise of foundation models just don't know the other side of how difficult it is to deploy these types of robots, he says.
Another issue is how far robot foundation models can get using the visual data that make up the vast majority of their physical training. Robots might need reams of other kinds of sensory data, for example from the sense of touch or proprioception a sense of where their body is in space say Soh. Those data sets dont yet exist. Theres all this stuff thats missing, which I think is required for things like a humanoid to work efficiently in the world, he says.
Releasing foundation models into the real world comes with another major challenge safety. In the two years since they started proliferating, large language models have been shown to come up with false and biased information. They can also be tricked into doing things that they are programmed not to do, such as telling users how to make a bomb. Giving AI systems a body brings these types of mistake and threat to the physical world. If a robot is wrong, it can actually physically harm you or break things or cause damage, says Gopalakrishnan.
Valuable work going on in AI safety will transfer to the world of robotics, says Gopalakrishnan. In addition, her team has imbued some robot AI models with rules that layer on top of their learning, such as not to even attempt tasks that involve interacting with people, animals or other living organisms. Until we have confidence in robots, we will need a lot of human supervision, she says.
Despite the risks, there is a lot of momentum in using AI to improve robots and using robots to improve AI. Gopalakrishnan thinks that hooking up AI brains to physical robots will improve the foundation models, for example giving them better spatial reasoning. Meta, says Rai, is among those pursuing the hypothesis that true intelligence can only emerge when an agent can interact with its world. That real-world interaction, some say, is what could take AI beyond learning patterns and making predictions, to truly understanding and reasoning about the world.
What the future holds depends on who you ask. Brooks says that robots will continue to improve and find new applications, but their eventual use is nowhere near as sexy as humanoids replacing human labour. But others think that developing a functional and safe humanoid robot that is capable of cooking dinner, running errands and folding the laundry is possible but could just cost hundreds of millions of dollars. Im sure someone will do it, says Khazatsky. Itll just be a lot of money, and time.
Originally posted here:
The AI revolution is coming to robots: how will it change them? - Nature.com
- How would a Victorian author write about generative AI? - Verdict [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Is the current regulatory system equipped to deal with AI? - The Hindu [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Tim Sweeney, CD Projekt, and Other Experts React to AI's Rise, and ... - IGN [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- GCHQ chiefs warning to ministers over risks of AI - The Independent [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Solving The Mystery Of How ChatGPT And Generative AI Can Surprisingly Pick Up Foreign Languages, Says AI Ethics And AI Law - Forbes [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- SenseAuto Empowers Nearly 30 Mass-produced Models Exhibited at Auto Shanghai 2023 and Unveils Six Intelligent Cabin Products - Yahoo Finance [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- AI Dangers Viewed Through the Perspective of Don't Look Up - BeInCrypto [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- ChatGPT, artificial intelligence, and the news - Columbia Journalism Review [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Unpacking AI: "an exponential disruption" with Kate Crawford: podcast and transcript - MSNBC [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Is 'Generative' AI the Way of the Future? - Northeastern University [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- GPT-4 Passes the Bar Exam: What That Means for Artificial ... - Stanford Law School [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Is artificial intelligence approaching science fiction? - The Trail - The Puget Sound Trail [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- Towards Artificial General Intelligence, ChatGPT 5 is on Track - Analytics Insight [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- New AGI hardware in progress for artificial general intelligence - Information Age [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- ChatGPT is impressive, but it may slow the emergence of AGI - TechTalks [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- How smart is ChatGPT really and how do we judge intelligence in AIs? - New Scientist [Last Updated On: August 18th, 2024] [Originally Added On: April 20th, 2023]
- AI Singapore and the Digital and Intelligence Service Sign ... - MINDEF Singapore [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- Artificial Intelligence Will Take Away Jobs and Disrupt Society, says Zerodha CEO Nithin Kamath - DATAQUEST [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- China's State-Sponsored AI Claims it Will Surpass ChatGPT by End ... - Tom's Hardware [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- AI At The Crossroads: Navigating Job Displacement, Ethical Concerns, And The Future Of Work - Forbes [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- I created a billion-pound start-up business Elon Musk & Jeff Bezos asked to meet me heres the secret to... - The Sun [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- Opinion | We Need a Manhattan Project for AI Safety - POLITICO [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- As AutoGPT released, should we be worried about AI? - Cosmos [Last Updated On: August 18th, 2024] [Originally Added On: May 14th, 2023]
- The Role of Artificial Intelligence in the Future of Media - Fagen wasanni [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- The Threat Of Climate Misinformation Propagated by Generative AI ... - Unite.AI [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- AI and the Next Phase of Human Evolution: What Can We Expect? - Fagen wasanni [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- IT pros mull observability tools, devx and generative AI - TechTarget [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- The future of learning and skilling with AI in the picture - Chief Learning Officer [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- The Economic Case for Generative AI and Foundation Models - Andreessen Horowitz [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Exploring the future of AI: The power of decentralization - Cointelegraph [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Convergence of Brain-Inspired AI and AGI: Exploring the Path to ... - Newswise [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Past, Present, Future: AI, Geopolitics, and the Global Economy - Tech Policy Press [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Development of GPT-5: The Next Step in AI Technology - Fagen wasanni [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Executive Q&A: Andrew Cardno, QCI - Indian Gaming [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- The Future is Now: Understanding and Harnessing Artificial ... - North Forty News [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- The Twin Convergence: AGI And Superconductors Ushering Humanity's Inflection Point - Medium [Last Updated On: August 4th, 2023] [Originally Added On: August 4th, 2023]
- Moonshot: Coexisting with AI holograms - The Edge Malaysia [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- A large computing cluster at sea might have big implications for AI ... - XDA Developers [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- ChatGPT or not ChatGPT? That was the question, briefly, as ... - GeekWire [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- The Best ChatGPT Prompts Are Highly Emotional, Study Confirms - Tech.co [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- The impact of AI and Language Models - Girton College [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Startup gaining investment traction for AI clinician productivity tool - Mobihealth News [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- What OpenAI's latest batch of chips says about the future of AI - Quartz [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- How AI Ecosystems Are Transforming the Future of Business - Entrepreneur [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Understanding Artificial Intelligence: Definition, Applications, and ... - Medium [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Europe's weaknesses, opportunities facing the AI revolution - EURACTIV [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- How the AI Executive Order and OMB memo introduce ... - Brookings Institution [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Artificial intelligence: the world is waking up to the risks - InCyber [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- How to win the artificial general intelligence race and not end ... - The Strategist [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- AI 2023: risks, regulation & an 'existential threat to humanity' - RTE.ie [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Game-playing DeepMind AI can beat top humans at chess, Go and poker - New Scientist [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- Sam Altman Seems to Imply That OpenAI Is Building God - Futurism [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- OpenAI's six-member board will decide 'when we've attained AGI' - VentureBeat [Last Updated On: August 18th, 2024] [Originally Added On: November 16th, 2023]
- The Most Important AI Innovations of 2024 | by AI News | Dec, 2023 - DataDrivenInvestor [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- Forget Dystopian Scenarios AI Is Pervasive Today, and the Risks Are Often Hidden - The Good Men Project [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- The Era of AI: 2023's Landmark Year - CMSWire [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- The Impact of OpenAIs GPT 5. A New Era of AI | by Courtney Hamilton | Dec, 2023 - Medium [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- What Is Artificial Intelligence? From Software to Hardware, What You Need to Know - ExtremeTech [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- AI consciousness: scientists say we urgently need answers - Nature.com [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- AI Technologies Set to Revolutionize Multiple Industries in Near Future - Game Is Hard [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- Will superintelligent AI sneak up on us? New study offers reassurance - Nature.com [Last Updated On: August 18th, 2024] [Originally Added On: December 27th, 2023]
- Amazon reportedly preparing paid Alexa version powered by its own Titan AI model - SiliconANGLE News [Last Updated On: August 18th, 2024] [Originally Added On: May 25th, 2024]
- OpenAI departures: Why cant former employees talk, but the new ChatGPT release can? - Vox.com [Last Updated On: August 18th, 2024] [Originally Added On: May 25th, 2024]
- Meta AI Head: ChatGPT Will Never Reach Human Intelligence - PYMNTS.com [Last Updated On: August 18th, 2024] [Originally Added On: May 25th, 2024]
- What is artificial general intelligence, and is it a useful concept? - New Scientist [Last Updated On: August 18th, 2024] [Originally Added On: May 25th, 2024]
- 22 jobs artificial general intelligence (AGI) may replace and 10 jobs it could create - Livescience.com [Last Updated On: August 18th, 2024] [Originally Added On: May 25th, 2024]
- Why AI Won't Take Over The World Anytime Soon - Bernard Marr [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]
- OpenAI announces new Safety and Security Committee as the AI race hots up and concerns grow around ethics - TechRadar [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]
- Responsible AI needs further collaboration - Chinadaily.com.cn - China Daily [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]
- OpenAI says it's charting a "path to AGI" with its next frontier AI model - ITPro [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]
- Can AI ever be smarter than humans? | Context - Context [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]
- OpenAI revives its robotic research team, plans to build dedicated AI - Interesting Engineering [Last Updated On: August 18th, 2024] [Originally Added On: June 3rd, 2024]