To accomplish this, we introduced the four-phase strategy listed below and discussed the first two phases, pre-processing and feature detection.

Preprocessing: Data is collected from the real world (e.g., from sensors and cameras) and converted into a more usable state.

Feature Detection: Features such as corners, edges, etc. are extracted from the preprocessed data.

Object Detection and Classification: Objects are detected from the features and may be classified according to known feature maps.

Object Tracking and Navigation: Objects that have been identified are tracked across time. This can include both objects and changing viewpoints of the environment as the robot navigates.

In this blog, we'll look at how the third and fourth phases, object detection and classification and object tracking and navigation, provide the foundation for higher-level robotic vision functionality.

Detecting Objects and Orientations

With features detected, the next step is to detect and classify objects from them. This has traditionally been a difficult task thanks to challenges like variations in viewpoints, different sized images, and varying illumination conditions. Fortunately, a neural network that has been trained to detect and classify objects with such variations can help. These networks are often trained from datasets containing large quantities of labelled images such as those illustrated here:

One popular approach is to employ a convolutional neural network (CNN), where small regions of the image are fed into the network in a process known as “sliding windows.” Developers can implement this using the Qualcomm® Neural Processing Engine SDK, which supports a wide variety of neural network layer types including convolution layers. The Qualcomm Neural Processing Engine SDK also supports output from Caffe2 and TensorFlow, as well as the ONNX format, so developers can take advantage of those frameworks in their object detection code.

Another task is to determine the orientation of objects, which is important for both object interaction and navigation. The main challenge here is determining the orientation of an object and/or the robot itself in 3D world-space. A popular approach is to apply homography algorithms such as linear least square solver, random sampling and consensus (RANSAC), and least median of squares, to compute points between frames of 2D imagery. The Qualcomm® Computer Vision library provides developers with hardware-accelerated homography and pose evaluation APIs for this purpose.

Once objects have been detected, they can then be assigned metadata such as an ID, bounding box, etc. which can be used during object detection and navigation.

Object Tracking and Navigation

With objects and aspects of the surrounding environment identified, a robot then needs to track them. Since objects can move around, and the robot's viewport will change as it navigates, developers will need a mechanism to track these elements over time and across frames captured by the camera(s) and other sensors. Since this mechanism must be fast enough to run every frame, numerous algorithms have been devised over the years which approach the problem in different ways.

For example, centroid tracking computes the center point of a bounding box around an identified object across frames, and computes the distance between the point as it changes, under the assumption that the object will only move a certain distance each frame. Another approach is to use a Kalman filter that uses statistics over time to predict the location of an object.

The Qualcomm Computer Vision library includes a hardware-accelerated implementation of the mean shift algorithm. This approach basically finds the mean of some aspect of an image (e.g., color histogram) within a sub region of a frame. It then looks for the same description within the next frame by seeking to maximize similarities in features. This allows it to account for changes such as scale, orientation etc. and to ultimately track where the object is. This is illustrated in the image below, where the algorithm is able to track the player's hand and pin-point its new location:

Since these techniques only need to track a subset of the original features, they can generally deal with changes such as orientation, occlusion etc., efficiently and with good success, which makes them effective for robotics vision processing.

But objects aren't the only thing that need to be tracked. The robot itself should be able to successfully navigate its environment and this is where Simultaneous Localization and Mapping (SLAM) comes in. SLAM seeks to estimate a robot's location and derive a map of the environment. It can be implemented using a number of algorithms such as Kalman filters. SLAM is often implemented by fusing data from multiple sensors, and when it involves visual data, the process is often referred to as Visual-Inertial Simultaneous Localization and Mapping (VISLAM).

Developers can use the Qualcomm® Machine Vision SDK, which supports a number of algorithms to determine position and orientation. This includes VISLAM using an extended Kalman filter that fuses camera and IMU data to derive a 6-DoF pose estimate in real world coordinates, depth from stereo, and voxel mapping.

Of course, SLAM is only as good as what the robot can sense, so developers should be sure to choose high-quality cameras and sensors, and find ways to ensure that they're not blocked from capturing data. From a safety aspect, developers should also devise fail safes in case data cannot be acquired (e.g. the cameras become covered).

Applying Computer Vision in Robotics

This four-phase strategy should provide a fairly rich set of data that developers can use as the basis of high-level robotic functionality. Developers who are interested in moving forward with next-generation robotics, can purchase the Qualcomm® Robotics RB3 development kit from Thundercomm and start taking advantage of the various SDKs available on Qualcomm Developer Network. And for those just getting started with robotics, be sure to check out our involvement with FIRST® (For Inspiration and Recognition of Science and Technology).

So now that you've seen a strategy for applying computer vision in robotics, we'd love to hear how you would approach your robotic development.

Related Tags:

Opinions expressed in the content posted here are the personal opinions of the original authors, and do not necessarily reflect those of Qualcomm Incorporated or its subsidiaries ("Qualcomm"). The content is provided for informational purposes only and is not meant to be an endorsement or representation by Qualcomm or any other party. This site may also provide links or references to non-Qualcomm sites and resources. Qualcomm makes no representations, warranties, or other commitments whatsoever about any non-Qualcomm sites or third-party resources that may be referenced, accessible from, or linked to this site.

Footer Links

References to "Qualcomm" may mean Qualcomm Incorporated, or subsidiaries or business units within the Qualcomm corporate structure, as applicable.

Qualcomm Incorporated includes Qualcomm's licensing business, QTL, and the vast majority of its patent portfolio. Qualcomm Technologies, Inc., a wholly-owned subsidiary of Qualcomm Incorporated, operates, along with its subsidiaries, substantially all of Qualcomm's engineering, research and development functions, and substantially all of its products and services businesses. Qualcomm products referenced on this page are products of Qualcomm Technologies, Inc. and/or its subsidiaries.

Materials that are as of a specific date, including but not limited to press releases, presentations, blog posts and webcasts, may have been superseded by subsequent events or disclosures.

Nothing in these materials is an offer to sell any of the components or devices referenced herein.