With robots being designed to handle increasingly complex and precise tasks, the fields of machine vision, image processing, and pattern recognition are gaining increasing importance. Researchers worldwide are now investigating promising robot vision technologies. Their goal is to creating robots that can easily navigate through various surroundings and recognize different types of objects with minimal or no human intervention.
Acoustic imaging reduces processor load
University researchers at ETH Zurich, for example, have created an acoustic imaging device that’s designed to show only the contours and edges of an object. This is an alternative to generating a more complex — and resource-draining — photorealistic image. According to project leader Chiara Daraio, an ETH professor of mechanics and materials, the new imaging technique is designed for when there is a need to quickly record critical information about an object rather than obtaining a fully detailed image.

A 3D-printed polymer structure with five resonance chambers. Microphones in the side holes on the left systematically scan a surface and generate an outline image from measured sound data
At the technology’s heart is a unique pipe-shaped polymer structure that’s produced on a 3D printer. The structure features a square cross-section interior that is divided into five adjoining resonance chambers linked via a series of small windows.
Robot vision mimics insects
Researchers at Australia’s University of Adelaide are taking a cue from the way insects see and track their prey to develop enhanced robot vision systems.
“Detecting and tracking small objects against complex backgrounds is a highly challenging task, observed research team member Zahra Bagheri, a mechanical engineering doctoral student. “Robotics engineers still dream of providing robots with the combination of sharp eyes, quick reflexes, and flexible muscles.”
Research conducted in the laboratory of Steven Wiederman, a neuroscientist at the University of Adelaide’s School of Medical Sciences, has shown that flying insects such as dragonflies exhibit a remarkable level of visually guided behavior when chasing mates or prey.
“They perform this task despite their low visual acuity and a tiny brain, around the size of a grain of rice,” Bagheri said. “The dragonfly chases prey at speeds up to 60 km/h, capturing them with a success rate over 97 percent.”

University of Adelaide Ph.D. student Zahra Bagheri and supervisor Prof. Benjamin Cazzolato with a mobile robot featuring a vision system using algorithms based on insect vision.
Robot vision gains multiple perspectives
Enabling household robots to recognize things faster and more accurately by imaging objects from multiple perspectives is the goal of researchers in the MIT Computer Science and Artificial Intelligence Laboratory. The researchers began their machine vision investigation by using a common algorithm that can combine different perspectives to recognize four times as many objects as an algorithm that uses a single perspective.
They then turned to a new algorithm that is just as accurate but can be up to 10 times as fast, making it much more practical for use in household robots.
“If you just took the output of looking at things from one viewpoint, there’s a lot of stuff that might be missing, or it might be the angle of illumination or something blocking the object that causes a systematic error in the detector,” said Lawson Wong, a graduate student in electrical engineering and computer science and lead author on the new paper. “One way around that is just to move around and go to a different viewpoint.”
Wong, working with thesis advisors Leslie Kaelbling, an computer science and engineering professor at the Massachusetts Institute of Technology, and TomAs Lozano-Perez, a professor of teaching excellence in the MIT School of Engineering, created scenarios in which 20 to 30 different images of household objects were placed near one another on a table. The first algorithm, developed years ago for various types of tracking systems, was used to analyze pairs of successive images and then create multiple hypotheses about which objects in one image correspond to objects in the other.
As new perspectives were added, the number of hypotheses rose. The drawback to the approach is that the algorithm must reject all but the most likely hypotheses at each step, a time-consuming task that is far from ideal for real-world robotic applications.
More on Robot Vision:
- AGVs Get an ETA With Seegrid’s Subway Platform
- Drone Funding Provides Lift for Specific Applications
- Seegrid’s IP Goldmine: Logistics, and Then Some
- Vicarious’ AI Vision Draws $70 Million From International Investors
- Australian Efforts to Hone Robotic Vision Get Government Support
- The Rise of Vision Systems Is a Turning Point for Robotics
Addressing this issue, the researchers developed an algorithm doesn’t reject any of the hypotheses generated over successive images. The algorithm also doesn’t attempt to fully evaluate the hypotheses, the process that’s primarily responsible for slowing down the algorithm’s final output.
Instead, the algorithm samples the hypotheses at random, taking advantage of the fact that there is a considerable overlap between multiple hypotheses. A sufficient number of samples, the researchers believe, should be sufficient to supply a consensus on the correspondences between the objects in any two successive images.
To keep the required number of samples low, the researchers adopted a simplified technique for evaluating hypotheses. In testing, the new algorithm reduced the number of matches generated by the first algorithm from 304 sets to only 20 comparisons.
On the downside, however, the shortcut could lead to the generation of nonsensical results created by the algorithm inadvertently mapping outcomes twice. To guard against false results, the new algorithm automatically searches for potential double mappings and re-evaluates them. The process demands additional time, yet the new algorithm is still far more efficient than its older counterpart. With the safety process in place, the algorithm performed 32 comparisons — more than 20, yet significantly less than 304.
Going deep for robot vision
Depth-sensing cameras, like the type included in the popular Microsoft Kinect video game controller, are widely used as 3D sensors in a variety of applications, including machine vision. A new imaging technology developed by researchers at the Carnegie Mellon University Robotics Institute and the University of Toronto Institute for Robotics and Mechatronics aims to resolve an important drawback found in such cameras: an inability to work in bright light, particularly sunlight.

A new depth-sensing camera technology developed by CMU and the University of Toronto can capture 3D information, such as this face, in full sunlight. Conventional depth cameras are typically blinded by bright light.
The researchers have created a mathematical model that helps the camera and its light source work together more efficiently, removing unwanted light that only serves to wash out the signals needed to detect an object’s contours.