Periodic Reporting for period 1 - REAL-RL (Model-based Reinforcement Learning for Versatile Robots in the Real World)
Okres sprawozdawczy: 2023-01-01 do 2025-06-30
REAL-RL takes a new approach that focuses on helping robots figure things out for themselves by trial and error. Most current methods in this area require robots to practice a huge number of times, which takes too long and can wear out real robots. As a shortcut, scientists use computer simulations, but these still need a lot of knowledge about every possible situation the robot might run into.
Our project is different: we want robots to learn their own "common sense" about how the world works, building up knowledge from their own experiences. With this knowledge, robots can quickly plan what to do and adapt when something unexpected happens—without needing tons of practice or super-detailed instructions.
We are tackling important scientific and technical challenges that stand in the way of this vision, including:
- How robots can build useful models of the world from experience.
- How they can safely explore and learn without making mistakes that could cause damage.
- How they can think ahead and make plans that take into account risks and different possible futures.
We’ve already made progress on some of the building blocks needed to make this happen, like computer programs that can guess what might happen next, or fast ways for robots to plan their actions.
Ultimately, if successful, this project will help build a new generation of robots that can use what they’ve learned to help humans with a wider range of tasks—from walking and carrying, to assembling items, to assisting around the house or workplace. These robots will not just follow instructions, but actually get better with practice, making them much more useful and adaptable in the future.
1. Helping robots understand how the world works.
We are teaching robots to build their own "mental models" of how they and their surroundings work, so they can predict what might happen next. For example, this could help a robot know how it will move when a certain force is applied. We tried using both traditional, physics-based tricks and newer, data-driven tricks—like learning directly from video footage. Early results are promising: we have working prototypes, and our research is being submitted to top scientific journals.
2. Making learning safer and more robust
For robots to learn without putting themselves or others at risk, we studied new ways to help them plan their actions more safely. We found that it’s not just about giving robots more data, but about giving them the right kind of data. We developed smarter ways for robots to learn from “what would have happened if…” scenarios, all while keeping them safe from risky mistakes. Several of our solutions are being published and shared with other researchers.
3. Teaching robots to make better decisions
We made improvements to how robots can plan ahead for tasks that involve complex decisions, like moving objects or navigating cluttered spaces.
Our new mathematical methods let computers "think ahead" during difficult choices, making planning much faster and more reliable. We have also begun exploring how large AI models, like those used in chatbots, can help robots reason through more complex situations.5. Powerful new tools
4. We are modifying an existing robot simulator to allow for a different way of obtaining control strategies for robots.
We’re also building a new type of simulator that learns directly from video, which could one day let robots “watch and learn” from what they see around them.
Our modifications to the robot simulator allows using them as flexible world models for mental simulation and allows for efficient controller generation. If successful, we are much closer to our goal of a versatile learning robot.