This post is also available in: עברית (Hebrew)
Researchers from Brown University have unveiled a transformative AI-based system that is expected to reshape how humans communicate with robots. This new and innovative system, called “Lang2LTL”, addresses the challenges in enabling robots to understand and carry out human instructions presented in everyday language.
Until now, trying to instruct robots using regular, everyday language and context-driven commands presented an impossible hurdle, which in turn necessitated extensive data-driven training for robots to decipher and execute nuanced instructions. However, recent breakthroughs in AI-driven large language models are now presenting a new era in human-robot communication.
Senior research author Stefanie Tellex explains that the team aimed to bridge the gap between complex human instructions and a robot’s actions, contemplating scenarios like guiding a mobile robot through nuanced paths or locations. “We wanted a way to connect complex, specific and abstract English instructions that people might say to a robot,” Tellex added, highlighting the system’s ability to interpret rich and precise instructions.
According to Interesting Engineering, the system excels in converting language directives into actionable robot behaviors without requiring huge amounts of training data. The system can seamlessly adapt to new environments without extensive training and merely requires a detailed map of the surroundings.
The researchers tested the system by conducting simulations across 21 cities using OpenStreetMap and achieved an impressive 80 percent accuracy rate, while existing systems typically perform a mere 20 percent accuracy. The system’s versatility makes it applicable to many different uses and scenarios, like guiding drones, self-driving cars, or ground vehicles through cityscapes, and overall facilitating intricate and precise instructions for navigating complex environments.
The system works by extracting locations from user instructions, matching them to known environments, and converting them into a format the robot comprehends, according to the study’s lead author Jason Xinyu. “Our system uses its modular system design and its large language models pre-trained on internet-scaled data to process more complex directional and linear-based natural language commands,” explained Xinyu.
Going forward, the researchers are expected to release a simulation allowing users to test the system’s functionality and provide valuable insights for further refinement. They also aim to integrate object manipulation capabilities into the software that would expand the system’s repertoire.
This revolutionary system heralds a new dawn in seamless, nuanced human-to-robot communication, promising widespread applications in many different domains.