Purdue Researcher Helps Robots 'See' in 3-D Like Humans

Figure-ground organization is key to seeing an object in 3-D instead of 2-D

WEST LAFAYETTE, Ind.--Zygmunt Pizlo and his research team glide across a parquet dance floor - not in some club for a night on the town, but in his Purdue University Visual Perception Lab as part of critical research for a technology that is ready to be licensed and commercialized.


They're moving so a robot named Čapek can "watch" them and conceptualize the research team's actions as members move around objects like desks and chairs. The goal is to simulate visual perception in the robot so it can "see" more like humans.

"Enabling robots and other machines to see the world in 3-D like humans is one of the biggest challenges in robotics and artificial intelligence," said Pizlo, a professor in the Purdue Department of Psychological Sciences. "Research in the field of robotic vision has typically focused on recording and analyzing 2-D images, but really it is about 3-D visual perception - being able to understand the 3-D scene in front of the robot so that it can decide what needs to be done with an object that is in its field of view. Should the robot walk around it? Pick it up?"

Pizlo has been working in the field of visual perception for 30 years.

"We believe there is a fundamental principle for human vision and that is we rely on a prior knowledge about a physical environment, so we're trying to program this knowledge of the physical environment into a robot's artificial intelligence," he said.

Postdoctoral research assistants Tadamasa Sawada and Yunfeng Li are working with Pizlo. Sawada said humans have the ability for cognitive functions that are computationally difficult, and it is a challenge to incorporate that ability into a robot.

"We quickly and easily perceive the physical world - a 3-D shape and figure-ground organization," Sawada said. "Figure-ground organization is key to seeing an object in 3-D instead of 2-D."

Figure-ground perception is the tendency of a human's visual system to simplify a scene or photo into a main object and cognitively "move" everything else into the background.

Conventional robotic vision technology uses multiple cameras with laser range finders and other sensors to detect objects around them. While the current systems allow for basic object recognition, they do not replicate the 3-D capabilities that are possible for humans, according to Pizlo.

Enabling the human/robot connection will be a key factor in bringing robots into everyday life.

"Right now robots are used in a number of ways, including manufacturing, space research, agriculture and even cleaning our floors, but they can't bring us coffee in the morning," he said. "Until they can see like us, they can't truly interact with us. Once they can interact with us they can begin doing all types of tasks such as drive a car, help surgeons in hospitals, assist the elderly, provide sight for the blind, replace people in high-risk situations like making repairs in a nuclear plant and, yes, bring us coffee in the morning."

A video about the project can be viewed at http://bit.ly/KmPGXy.

The patent-pending technology is available for licensing through Eric Lynch at 765-588-3477, eslynch@prf.org, in the Purdue Office of Technology Commercialization.

Funding for Pizlo's research came from the National Science Foundation, U.S. Department of Defense, Air Force Office of Scientic Research, U.S. Department of Energy and other sources.

Featured Product

Robotmaster® 2024

Robotmaster® 2024

Program multi-robot cells and automatically solve robotic errors with ease. Hypertherm Associates announces a new version to its robotic programming software. Robotmaster 2024 addresses key market trends including the support for programming multiple robots in a single work cell and the demand for automatic trajectory optimization and robotic error correction.