Cornell engineers train 'checkout chick' robot

Thursday, 07 November, 2013

Are humans becoming obsolete? The advent of the self-checkout at our supermarkets might lead one to think so, despite the fact that they are invariably overseen by one of our species - more to help the humans deal with the computers than the other way around.

In a further step towards automation in the supermarket, Cornell University engineers have trained a robot to work as a ‘checkout chick’ in a simulated grocery checkout.

The robot, a Baxter robot from Rethink Robotics in Boston, was modified to ‘coactively learn’ from humans and make adjustments while an action is in progress.

“We give the robot a lot of flexibility in learning,” said Ashutosh Saxena, Assistant Professor of Computer Science at Cornell. “The robot can learn from corrective human feedback in order to plan its actions that are suitable to the environment and the objects present.”

Modern industrial robots, like those on automobile assembly lines, have no brains, just memory. An operator programs the robot to move through the desired action; the robot can then repeat the exact same action every time an object goes by.

But off the assembly line, things get complicated: a personal robot working in a home has to handle tomatoes more gently than canned goods. If it needs to pick up and use a sharp kitchen knife, it should be smart enough to keep the blade away from humans.

The Baxter’s arms have two elbows and a rotating wrist, so it’s not always obvious to a human operator how best to move the arms to accomplish a particular task. So Saxena and graduate student Ashesh Jain drew on previous work, adding programming that lets the robot plan its own motions. It displays three possible trajectories on a touch screen where the operator can select the one that looks best.

Then humans can give corrective feedback. As the robot executes its movements, the operator can intervene, manually guiding the arms to finetune the trajectory. The robot has what the researchers call a “zero-G” mode, where the robot’s arms hold their position against gravity but allow the operator to move them. The first correction may not be the best one, but it may be slightly better. The learning algorithm the researchers provided allows the robot to learn incrementally, refining its trajectory a little more each time the human operator makes adjustments or selects a trajectory on the touch screen. Even with weak but incrementally correct feedback from the user, the robot arrives at an optimal movement.

The robot learns to associate a particular trajectory with each type of object. A quick flipover might be the fastest way to move a cereal box, but that wouldn’t work with a carton of eggs. Also, since eggs are fragile, the robot is taught that they shouldn’t be lifted far above the counter. Likewise, the robot learns that sharp objects shouldn’t be moved in a wide swing; they are held in close, away from people.

In tests with users who were not part of the research team, most users were able to train the robot successfully on a particular task with just five corrective feedbacks. The robots also were able to generalise what they learned, adjusting when the object, the environment or both were changed.

Saxena’s research team will report their work at the Neural Information Processing Systems conference in Lake Tahoe, California, on 5 to 8 December 2013.

Related News

Major US defence company sets up in SA

US defence technology company Sierra Nevada Corporation has opened an Australian subsidiary, SNC...

Queensland boosting local manufacturing

The Queensland Government has announced it will boost homegrown manufacturing with an expansion...

Hazer completes testing of its commercial demonstration plant

Hazer Group has announced that it has completed testing of its commercial demonstration plant for...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd