- It can be extremely dangerous to send in first responders, during hurricanes, flash flooding and other disasters.
- A research team have been exploring how autonomous drones could help identify people in need and notify rescue teams.
- Many rescuers already use drones, but most require individual pilots who fly the unmanned aircraft by remote control, slowing the process.
- Autonomous drones can scan the landscape, investigate the selected region, detect humans and collects GPS data in about one-fifth of a second.
During hurricanes, flash flooding and other disasters, it can be extremely dangerous to send in first responders, even though people may badly need help.
Rescuers already use drones in some cases, but most require individual pilots who fly the unmanned aircraft by remote control. That limits how quickly rescuers can view an entire affected area, and it can delay aid from reaching victims.
Autonomous drones could cover more ground faster, especially if they could identify people in need and notify rescue teams.
My team and I at the University of Dayton Vision Lab have been designing these autonomous systems of the future to eventually help spot people who might be trapped by debris. Our multi-sensor technology mimics the behavior of human rescuers to look deeply at wide areas and quickly choose specific regions to focus on, examine more closely, and determine if anyone needs help.
The deep learning technology that we use mimics the structure and behavior of a human brain in processing the images captured by the 2-dimensional and 3D sensors embedded in the drones. It is able to process large amounts of data simultaneously to make decisions in real time.
Looking for an object in a chaotic scene
Disaster areas are often cluttered with downed trees, collapsed buildings, torn-up roads and other disarray that can make spotting victims in need of rescue very difficult. 3D lidar sensor technology, which uses light pulses, can detect objects hidden by overhanging trees.
My research team developed an artificial neural network system that could run in a computer onboard a drone. This system emulates some of the ways human vision works. It analyzes images captured by the drone’s sensors and communicates notable findings to human supervisors.
First, the system processes the images to improve their clarity. Just as humans squint their eyes to adjust their focus, this technology take detailed estimates of darker regions in a scene and computationally lightens the images.
In a rainy environment, human brains use a brilliant strategy to see clearly: By noticing the parts of a scene that don’t change as the raindrops fall, people can see reasonably well despite the rain. Our technology uses the same strategy, continuously investigating the contents of each location in a sequence of images to get clear information about the objects in that location.
Have you read?
Confirming objects of interest
When rescuers search for human beings trapped in disaster areas, the viewers’ minds imagine 3D views of how a person might appear in the scene. They should be able to detect the presence of a trapped human even if they haven’t seen someone in such a position before.
We employ this strategy by computing 3D models of people and rotating the shapes in all directions. We train the autonomous machine to perform exactly like a human rescuer does. That allows the system to identify people in various positions, such as lying prone or curled in the fetal position, even from different viewing angles and in varying lighting and weather conditions.
The system can also be trained to detect and locate a leg sticking out from under rubble, a hand waving at a distance, or a head popping up above a pile of wooden blocks. It can tell a person or animal apart from a tree, bush or vehicle.
Putting the pieces together
During its initial scan of the landscape, the system mimics the approach of an airborne spotter, examining the ground to find possible objects of interest or regions worth further examination, and then looking more closely. For example, an aircraft pilot who is looking for a truck on the ground would typically pay less attention to lakes, ponds, farm fields and playgrounds because trucks are less likely to be in those areas. The autonomous technology employs the same strategy to focus the search area to the most significant regions in the scene.
Then the system investigates each selected region to obtain information about the shape, structure and texture of objects there. When it detects a set of features that matches a human being or part of a human, it flags that location, collects GPS data and senses how far the person is from other objects to provide an exact location.
The entire process takes about one-fifth of a second.
How has the World Economic Forum helped initiate a more effective response to natural disasters and humanitarian crises?
In 2005, the World Economic Forum helped to establish the Logistics Emergency Teams (LET), a network of representatives from four of the world’s largest logistics and transport companies (Agility, DP World, Maersk and UPS) who work together in partnership with the World Food Programme-led Global Logistics Cluster to deliver free humanitarian assistance.
To date, the LET has responded to more than 20 large-scale natural disasters and humanitarian crises, providing critical logistical support for hurricane victims in Haiti, Rohingya refugees in Bangladesh, tsunami victims in Indonesia, civilians in war-ravaged Yemen and many more.
In 2018, 1,943 employees of LET member companies were trained in humanitarian logistics, contingency operations and disaster response to ensure that they were better prepared for future crises.
Read more about how the LET initiative continues to be an exemplary model for public-private partnerships.
Contact us if you're interested in getting involved in impactful initiatives as a member or partner of the World Economic Forum.
This is what faster search-and-rescue operations can look like in the future. A next step will be to turn this technology into an integrated system that can be deployed for emergency response.
We previously worked with the U.S. Army Medical Research and Materiel Command on technology to find wounded individuals in a battlefield who need rescue. We also have adapted the technology to help utility companies monitor heavy equipment that could damage pipelines. These are just a few of the ways disaster responders, companies or even farmers could benefit from technology that can see as humans can see, especially in places humans can’t easily reach.