[ad_1]
From wiping up spills to serving up food, robots are being taught to carry out increasingly complicated household tasks. Many such home-bot trainees are learning through imitation; they are programmed to copy the motions that a human physically guides them through.
It turns out that robots are excellent mimics. But unless engineers also program them to adjust to every possible bump and nudge, robots donโt necessarily know how to handle these situations, short of starting their task from the top.
Now MIT engineers are aiming to give robots a bit of common sense when faced with situations that push them off their trained path. Theyโve developed a method that connects robot motion data with the โcommon sense knowledgeโ of large language models, or LLMs.
Their approach enables a robot to logically parse many given household task into subtasks, and to physically adjust to disruptions within a subtask so that the robot can move on without having to go back and start a task from scratch โ and without engineers having to explicitly program fixes for every possible failure along the way. ย ย
โImitation learning is a mainstream approach enabling household robots. But if a robot is blindly mimicking a humanโs motion trajectories, tiny errors can accumulate and eventually derail the rest of the execution,โ says Yanwei Wang, a graduate student in MITโs Department of Electrical Engineering and Computer Science (EECS). โWith our method, a robot can self-correct execution errors and improve overall task success.โ
Wang and his colleagues detail their new approach in a study they will present at the International Conference on Learning Representations (ICLR) in May. The studyโs co-authors include EECS graduate students Tsun-Hsuan Wang and Jiayuan Mao, Michael Hagenow, a postdoc in MITโs Department of Aeronautics and Astronautics (AeroAstro), and Julie Shah, the H.N. Slater Professor in Aeronautics and Astronautics at MIT.
Language task
The researchers illustrate their new approach with a simple chore: scooping marbles from one bowl and pouring them into another. To accomplish this task, engineers would typically move a robot through the motions of scooping and pouring โ all in one fluid trajectory. They might do this multiple times, to give the robot a number of human demonstrations to mimic.
โBut the human demonstration is one long, continuous trajectory,โ Wang says.
The team realized that, while a human might demonstrate a single task in one go, that task depends on a sequence of subtasks, or trajectories. For instance, the robot has to first reach into a bowl before it can scoop, and it must scoop up marbles before moving to the empty bowl, and so forth. If a robot is pushed or nudged to make a mistake during any of these subtasks, its only recourse is to stop and start from the beginning, unless engineers were to explicitly label each subtask and program or collect new demonstrations for the robot to recover from the said failure, to enable a robot to self-correct in the moment.
โThat level of planning is very tedious,โ Wang says.
Instead, he and his colleagues found some of this work could be done automatically by LLMs. These deep learning models process immense libraries of text, which they use to establish connections between words, sentences, and paragraphs. Through these connections, an LLM can then generate new sentences based on what it has learned about the kind of word that is likely to follow the last.
For their part, the researchers found that in addition to sentences and paragraphs, an LLM can be prompted to produce a logical list of subtasks that would be involved in a given task. For instance, if queried to list the actions involved in scooping marbles from one bowl into another, an LLM might produce a sequence of verbs such as โreach,โ โscoop,โ โtransport,โ and โpour.โ
โLLMs have a way to tell you how to do each step of a task, in natural language. A humanโs continuous demonstration is the embodiment of those steps, in physical space,โ Wang says. โAnd we wanted to connect the two, so that a robot would automatically know what stage it is in a task, and be able to replan and recover on its own.โ
Mapping marbles
For their new approach, the team developed an algorithm to automatically connect an LLMโs natural language label for a particular subtask with a robotโs position in physical space or an image that encodes the robot state. Mapping a robotโs physical coordinates, or an image of the robot state, to a natural language label is known as โgrounding.โ The teamโs new algorithm is designed to learn a grounding โclassifier,โ meaning that it learns to automatically identify what semantic subtask a robot is in โ for example, โreachโ versus โscoopโ โ given its physical coordinates or an image view.
โThe grounding classifier facilitates this dialogue between what the robot is doing in the physical space and what the LLM knows about the subtasks, and the constraints you have to pay attention to within each subtask,โ Wang explains.
The team demonstrated the approach in experiments with a robotic arm that they trained on a marble-scooping task. Experimenters trained the robot by physically guiding it through the task of first reaching into a bowl, scooping up marbles, transporting them over an empty bowl, and pouring them in. After a few demonstrations, the team then used a pretrained LLM and asked the model to list the steps involved in scooping marbles from one bowl to another. The researchers then used their new algorithm to connect the LLMโs defined subtasks with the robotโs motion trajectory data. The algorithm automatically learned to map the robotโs physical coordinates in the trajectories and the corresponding image view to a given subtask.
The team then let the robot carry out the scooping task on its own, using the newly learned grounding classifiers. As the robot moved through the steps of the task, the experimenters pushed and nudged the bot off its path, and knocked marbles off its spoon at various points. Rather than stop and start from the beginning again, or continue blindly with no marbles on its spoon, the bot was able to self-correct, and completed each subtask before moving on to the next. (For instance, it would make sure that it successfully scooped marbles before transporting them to the empty bowl.)
โWith our method, when the robot is making mistakes, we donโt need to ask humans to program or give extra demonstrations of how to recover from failures,โ Wang says. โThatโs super exciting because thereโs a huge effort now toward training household robots with data collected on teleoperation systems. Our algorithm can now convert that training data into robust robot behavior that can do complex tasks, despite external perturbations.โ
[ad_2]
Source link