Emerging Technologies

This robot can understand liquids and pour a glass of water, here's how

Robots trained to pour water could become servers, measure or mix medicines, or water gardens.

Robots trained to pour water could become servers, measure or mix medicines, or water gardens. Image: Unsplash/Jamie Street

Camden Flath
Author, Futurity
Share:
Our Impact
What's the World Economic Forum doing to accelerate action on Emerging Technologies?
The Big Picture
Explore and monitor how Water is affecting economies, industries and global issues
A hand holding a looking glass by a lake
Crowdsource Innovation
Get involved with our crowdsourced digital platform to deliver impact at scale
Stay up to date:

Emerging Technologies

  • Researchers have used AI and image translation to teach a robot to recognize water and pour it into a glass.
  • Robots struggle to understand transparent liquids because of the many ways they reflect, refract and absorb light.
  • By playing videos behind a transparent glass of water, the robot is trained to pour water against varied backgrounds in the real world, regardless of where it is located.
  • This discovery could enable robot servers to refill water glasses, robot pharmacists to measure and mix medicines, or robot gardeners to water plants.

A horse, a zebra, and artificial intelligence helped researchers teach a robot to recognize water and pour it into a glass.

Water presents a tricky challenge for robots because it is clear. Robots have learned how to pour water before, but previous techniques like heating the water and using a thermal camera or placing the glass in front of a checkerboard background don’t transition well to everyday life.

An easier solution could enable robot servers to refill water glasses, robot pharmacists to measure and mix medicines, or robot gardeners to water plants.

Now, researchers have used AI and image translation to solve the problem.

Gif of robot pouring water
Water presents a tricky challenge for robots. Image: Futurity

Teaching a robot to pour water

Image translation algorithms use collections of images to train artificial intelligence to convert images from one style to another, such as transforming a photo into a Monet-style painting or making an image of a horse look like a zebra. For this research, the team used a method called contrastive learning for unpaired image-to-image translation, or CUT, for short.

“You need some way of telling the algorithm what the right and wrong answers are during the training phase of learning,” says David Held, an assistant professor in the Robotics Institute at Carnegie Mellon University. “However, labeling data can be a time-consuming process, especially for teaching a robot to pour water, for which the human might need to label individual water droplets in an image.”

Enter the horse and zebra.

“Just like we can train a model to translate an image of a horse to look like a zebra, we can similarly train a model to translate an image of colored liquid into an image of transparent liquid,” Held says. “We used this model to enable the robot to understand transparent liquids.”

A transparent liquid like water is hard for a robot to see because the way it reflects, refracts, and absorbs light varies depending on the background. To teach the computer how to see different backgrounds through a glass of water, the team played YouTube videos behind a transparent glass full of water. Training the system this way will allow the robot to pour water against varied backgrounds in the real world, regardless of where the robot is located.

“Even for humans, sometimes it’s hard to precisely identify the boundary between water and air,” says Gautham Narasimhan, who earned his master’s degree in the Robotics Institute in 2020 and worked with a team in the institute’s Robots Perceiving and Doing Lab on the new work.

Using their method, the robot was able to pour the water until it reached a certain height in a glass. They then repeated the experiment with glasses of different shapes and sizes.

Narasimhan says there’s room for future research to expand upon this method—adding different lighting conditions, challenging the robot to pour water from one container to another, or estimating not only the height of the water, but also the volume.

The researchers presented their work at the IEEE International Conference on Robotics and Automation in May 2022.

“People in robotics really appreciate it when research works in the real world and not just in simulation,” says Narasimhan, who now works as a computer vision engineer with Path Robotics in Columbus, Ohio. “We wanted to do something that’s quite simple yet effective.”

Discover

How is the World Economic Forum ensuring the responsible use of technology?

Have you read?
Loading...
Don't miss any update on this topic

Create a free account and access your personalized content collection with our latest publications and analyses.

Sign up for free

License and Republishing

World Economic Forum articles may be republished in accordance with the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International Public License, and in accordance with our Terms of Use.

The views expressed in this article are those of the author alone and not the World Economic Forum.

Share:
World Economic Forum logo
Global Agenda

The Agenda Weekly

A weekly update of the most important issues driving the global agenda

Subscribe today

You can unsubscribe at any time using the link in our emails. For more details, review our privacy policy.

Stanford just released its annual AI Index report. Here's what it reveals

James Fell

April 26, 2024

About Us

Events

Media

Partners & Members

  • Join Us

Language Editions

Privacy Policy & Terms of Service

© 2024 World Economic Forum