But that places a significant limitation on drone operations. The whole point of drones is that they are unmanned. Without a human operator on board, though, how can a drone steer clear of collisions? This is a crucial problem for Amazon, Google, and any other company that wants to deliver packages with drones.

To be practical, delivery drones would have to be able to fly long distances, well out of sight of a human operator. How, then, can the operator prevent the drone from hitting a tree, building, airplane, or even another drone? Although cameras could be mounted on the drone for this purpose, current civil drone video transmission technology is limited to a range of a few miles. As a result, in order to perform long-distance deliveries, the drone must autonomously detect nearby objects and avoid hitting them.

As a drone operations researcher, I keep a close eye on ways to achieve this. New research into sensors – at least some of which come from development of autonomous cars – is making increased autonomy possible for drones, potentially opening the skies to even more innovation.

Article continues below

The revolution will not be televised. It'll be sent to your inbox by us.

Lidar, developed more recently, uses laser beams instead of radio waves, and can provide extremely detailed images of nearby features. The catch is that both radar and lidar systems have been bulky, heavy, and expensive. That makes them hard to fit on relatively small drones; also, heavier drones require more battery power to stay aloft, which requires bigger (and heavier) batteries.

A small lidar sensor.

There is hope, though. Research in obstacle sensors and collision avoidance technology for autonomous automobiles has spurred the development of small, lower-cost radar and lidar devices. Once they are sufficiently small, and energy-efficient enough not to quickly drain drone batteries, both types of sensors could help solve the drone “see-and-avoid,” or really, because drones don’t have eyes, the “detect-and-avoid” problem.

An In-Flight View

A recent test flight here at Ohio University involved a lidar sensor mounted on a drone. When the drone was approximately five feet above the ground, the lidar was able to create an image of its surroundings.

A lidar image from a drone in flight.

On one side, the image had bushy-looking areas representing trees and foliage. One the other, there were parallel lines indicating the location of a building wall. And in the middle, were some circular shapes representing the ground. This sort of obstacle detection capability and discernment will be essential for routine drone operation, particularly during takeoff and landing.

We are currently in what might be called the “Wright Brothers era” of drone development. Removing the human from the cockpit has challenged innovators and designers in a number of ways – including solving the task of obstacle detection. But as our technology advances, eventually – just like elevators that used to be operated by humans – people will grow used to the idea of these machines operating autonomously.