In this year, lockdowns and other COVID-19 safety measures have made online shopping more popular than ever, but the skyrocketing demand is leaving many retailers struggling to fulfill orders while ensuring the safety of their warehouse employees.
Researchers at the University of California, Berkeley, have created new artificial intelligence software that gives robots the speed and skill to grasp and smoothly move objects, making it feasible for them to soon assist humans in warehouse environments. The technology is described in a paper published online in the journal Science Robotics.
Automating warehouse tasks can be challenging because many actions that come naturally to humans — like deciding where and how to pick up different types of objects and then coordinating the shoulder, arm and wrist movements needed to move each object from one location to another — are actually quite difficult for robots. Robotic motion also tends to be jerky, which can increase the risk of damaging both the products and the robots.
Warehouses are still operated primarily by humans, because it’s still very hard for robots to reliably grasp many different objects.
In an automobile assembly line, the same motion is repeated over and over again, so that it can be automated. But in a warehouse, every order is different.
In earlier work, a UC Berkeley researcher created a Grasp-Optimized Motion Planner that could compute both how a robot should pick up an object and how it should move to transfer the object from one location to another.
However, the motions generated by this planner were jerky. While the parameters of the software could be tweaked to generate smoother motions, these calculations took an average of about half a minute to compute.
In the new study, the UC Berkeley researchers , dramatically sped up the computing time of the motion planner by integrating a deep learning neural network.
Neural networks allow a robot to learn from examples. Later, the robot can often generalize to similar objects and motions.
However, these approximations aren’t always accurate enough. The researchers found that the approximation generated by the neural network could then be optimized using the motion planner.
The neural network takes only a few milliseconds to compute an approximate motion. It’s very fast, but it’s inaccurate. However, if we then feed that approximation into the motion planner, the motion planner only needs a few iterations to compute the final motion.
By combining the neural network with the motion planner, the team cut average computation time from 29 seconds to 80 milliseconds, or less than one-tenth of a second.
With this and other advances in robotic technology, robots could be assisting in warehouse environments in the next few years.
News source: University of California, Berkeley