Skip to content

Latest commit

 

History

History
46 lines (34 loc) · 6.08 KB

README.md

File metadata and controls

46 lines (34 loc) · 6.08 KB

PSIAP-CAP-Annotation

The Civil Air Patrol (CAP) Analysis is a web application, primarily built using HTML CSS, Node.js, and Python 3, to crowdsource annotations of freely and publicly available CAP imagery. CAP is a congressionally chartered, federally supported non-profit corporation that serves as the official civilian auxiliary of the United States Air Force. They provide imagery, often collected at low altitudes, of incidients and disasters.

Overview and Motivation

This application, funded by PSIAP 2017, supports the development of datasets representative of public safety operations, whose need was formally identified in NIST TN 1917: Public Safety Analytics R&D Roadmap:

One of the most fundamental barriers to seamless data integration is simply a lack of awareness or access to datasets that are accurate, current, and relevant to improving response.

This effort is motivation by video and imagery becoming increasingly important to public safety operations. The goal of annotating CAP imagery is to accelerate technology innovation for network providers, application providers, and public safety agenices. Public safety outreach indicated the need for a low altitude airborne-based dataset. There is a desire and intent to deploy drones and low altitude aircraft in response to small and large incidents. These operations often occur at the tactical edge in disconnected, intermittent, and low-bandwidth (DIL) environments. Analytics trained on CAP imagery could improve situational awareness and result in a faster, safer, and more efficient response.

Development of any dataset for public safety is a large combinatorial challenge, as incidents and disasters can widely vary. CAP imagery is ideal because they respond to a wide range of incidents and the data is intended for public access and use. Most importantly, as CAP is a national resource that operates independent on the PSIAP effort and will continue to collect and post imagery beyond the PSIAP effort. A CAP-based dataset is every growing. This dynamism is a realization discussed in NISTIR 8164: First Workshop on Video Analytics in Public Safety:

Thoughts on future development: Increasingly, software developers will need to take advantage of new hardware innovations. With the increasing reliance on machine learning methods such as deep learning, developers will require access to ever-increasing quantities of data for both training and evaluation purposes. Going beyond static datasets (which can be overlearned), future algorithms will require constant novelty allowing for a state of never-ending unsupervised learning. This will require a migration from data-sets to data-sites and, perhaps, data-cities.

Prior to this PSIAP effort, current annotation capabilities for CAP imagery has good intent but are insufficient for response and research. Other efforts have focused on damage to buildings or structures, whereas public safety requires additional information. For example, an urban search and rescue (USAR) team would like situational awareness of what roads are flooded or bridges destroyed, so they can efficiently plan their search. The PSIAP effort intends to make the CAP annotations freely and publicly available.

Semantic Annotations

Liu et al. demonstrated that labeling individual features with bounding boxes is not required to produce useful annotations and analytics. Instead each CAP image is labeled with semantic text labels, such as bridge or vegetation. Multiple labels can easily and rapidly be assigned to faciliate efficient crowd sourcing. Labels can then be encoded as binary values to streamline their use as a training dataset for machine learning. Images are segmented and labeled in quadrants to improve the information for each image and better utilize the smaller screens of mobile devices.

Example CAP Imagery

Debris Flooding Damage

Useful Links

Authors

  • Daniel Ribeirinha-Braga (MITLL)

Acknowledgments

  • Andrew Weinert (MITLL)
  • Dieter Schuldt (MITLL)
  • Gabriela Barrera (MITLL)
  • Chad Council (MITLL)
  • Steven Talpas (NJOHSP)
  • William Drew (NJOHSP)

Distribution Statement

This work was performed under the following financial assistance award 70NANB17Hl69 from U.S. Department of Commerce, National Institute of Standards and Technology.