The world’s first drone deliveries have begun trial runs in the United Kingdom and the U.S. Once primarily used by militaries, small quadcopter and octocopter drones are now so commonplace they are for sale at home improvement stores and toy stores. People are flying drones for fun, for entertainment and for commercial purposes as diverse as filmmaking and farming.
All these uses have one thing in common: The drone’s human operator is required by law to be able to see the drone at all times. Why? The answer is simple: to make sure the drone doesn’t hit anything.
Beyond just wanting not to crash and damage their drones or themselves, drone operators must avoid collisions with people, property and other vehicles. Specifically, federal aviation regulations forbid aircraft – including drones – from flying “so close to another aircraft as to create a collision hazard.” The rules also require that “vigilance shall be maintained by each person operating an aircraft so as to see and avoid other aircraft.” These requirements are commonly referred to simply as “see-and-avoid”: Pilots must see and avoid other traffic.
But that places a significant limitation on drone operations. The whole point of drones is that they are unmanned. Without a human operator on board, though, how can a drone steer clear of collisions? This is a crucial problem for Amazon, Google and any other company that wants to deliver packages with drones.
To be practical, delivery drones would have to be able to fly long distances, well out of sight of a human operator. How, then, can the operator prevent the drone from hitting a tree, building, airplane or even another drone? Although cameras could be mounted on the drone for this purpose, current civil drone video transmission technology is limited to a range of a few miles. As a result, in order to perform long-distance deliveries, the drone must autonomously detect nearby objects and avoid hitting them.
As a drone operations researcher, I keep a close eye on ways to achieve this. New research into sensors – at least some of which come from development of autonomous cars – is making increased autonomy possible for drones, potentially opening the skies to even more innovation.
Radar and lidar
There are two main technologies available for drones to detect nearby objects. The first is radar, developed just before World War II, that sends out radio waves and measures their reflections from obstacles. Radar is still used as the primary system for air traffic controllers to track planes in the sky. Ships also use radar to avoid collisions at night or in foggy conditions.
Lidar, developed more recently, uses laser beams instead of radio waves, and can provide extremely detailed images of nearby features. The catch is that both radar and lidar systems have been bulky, heavy and expensive. That makes them hard to fit on relatively small drones; also, heavier drones require more battery power to stay aloft, which requires bigger (and heavier) batteries.

A small lidar sensor. Velodyne, CC BY-ND
There is hope, though. Research in obstacle sensors and collision avoidance technology for autonomous automobiles has spurred the development of small, lower-cost radar and lidar devices. Once they are sufficiently small, and energy-efficient enough not to quickly drain drone batteries, both types of sensors could help solve the drone “see-and-avoid,” or really, because drones don’t have eyes, the “detect-and-avoid” problem.
An in-flight view
A recent test flight here at Ohio University involved a lidar sensor mounted on a drone. When the drone was approximately five feet above the ground, the lidar was able to create an image of its surroundings.
A lidar image from a drone in flight. Michael Braasch, CC BY-ND
On one side, the image had bushy-looking areas representing trees and foliage. One the other there were parallel lines indicating the location of a building wall. And in the middle were some circular shapes representing the ground. This sort of obstacle detection capability and discernment will be essential for routine drone operation, particularly during takeoff and landing.
We are currently in what might be called the “Wright Brothers era” of drone development. Removing the human from the cockpit has challenged innovators and designers in a number of ways – including solving the task of obstacle detection. But as our technology advances, eventually – just like elevators that used to be operated by humans – people will grow used to the idea of these machines operating autonomously.
Michael Braasch has received funding from the FAA and NASA for drone research.
This article was originally published on The Conversation. Read the original article.


Google's TurboQuant Algorithm Sends Memory Chip Stocks Tumbling
Makemation: a Nollywood movie that shows AI in action in Africa
OpenAI Pulls the Plug on Sora, Ending $1 Billion Disney Partnership
SpaceX IPO Filing Expected This Week as Valuation Could Surpass $75 Billion
California's AI Executive Order Pushes Responsible Tech Use in State Contracts
Reflection AI Eyes $25 Billion Valuation in Massive $2.5 Billion Funding Round
SK Hynix Eyes Up to $14 Billion U.S. IPO to Fund AI Chip Expansion
Cybersecurity Stocks Tumble After Anthropic's Claude Mythos AI Leak Sparks Market Fears
MATCH Act Targets ASML and Chinese Chipmakers in New U.S. Export Crackdown
NASA's Artemis II Crew Arrives in Florida for Historic Moon Mission
Microsoft Eyes $7B Texas Energy Deal to Power AI Data Centers
Australia's Social Media Ban for Under-16s Sparks Global Movement
Microsoft's $10 Billion Japan Investment: AI Infrastructure and Data Sovereignty Push
Meta Ties Executive Pay to Aggressive Stock Price Targets in Major Retention Push
Nanya Technology Shares Surge 10% After $2.5 Billion Private Placement from Sandisk and Cisco
TSMC Japan's Second Fab to Produce 3nm Chips by 2028 




