Visual Cloud Can Aid First Responders

Tuesday, June 28, 2016 @ 12:06 PM gHale


In any kind of disaster, the ability to process massive amounts of visual electronic data quickly and efficiently could mean the difference between life and death for survivors.

Visual data created by numerous security cameras, personal mobile devices and aerial video provide useful data for first responders and law enforcement.

RELATED STORIES
Recycling Plastic Waste into Fuel
Electronic Bacteria Sensor for Food Safety
Making Nuclear Fuel Recycling Cleaner
Ultra-Sensitive Gas Sensor

That data can be critical in terms of knowing where to send emergency personnel and resources, tracking suspects in man-made disasters, or detecting hazardous materials.

Along those lines, a visual cloud computing architecture that streamlines the process.

“In disaster scenarios, the amount of visual data generated can create a bottleneck in the network,” said Prasad Calyam, assistant professor of computer science in the University of Missouri College of Engineering. “This abundance of visual data, especially high-resolution video streams, is difficult to process even under normal circumstances. In a disaster situation, the computing and networking resources needed to process it may be scarce and even not be available. We are working to develop the most efficient way to process data and study how to quickly present visual information to first responders and law enforcement.”

The research team, including Kannappan Palaniappan and Ye Duan, associate professors in the Department of Computer Science, developed a framework for disaster incident data computation that links the system to mobile devices in a mobile cloud.

Algorithms designed by the team help determine what information needs to be processed by the cloud and what information can end up processed on local devices, such as laptops and smartphones. This spreads the processing over multiple devices and helps responders receive the information faster.

“Often, we see many of the same images from overlapping cameras,” Palaniappan said. “Responders generally do not need to see two separate pictures but rather the distinctive parts. That mosaic stitching that we helped define happens in the periphery of the network to limit the amount of data that needs to be sent to the cloud. This is a natural way of compressing visual data without losing information. Clever algorithms help determine what types of visual processing to perform in the edge or fog of the network, and what data and computation should be done in the core cloud.”