Share Email Print
cover

Proceedings Paper • new

Augmentation techniques for video surveillance in the visible and thermal spectral range
Author(s): Vanessa Buhrmester; Ann-Kristin Grosselfinger; David Münch; Michael Arens
Format Member Price Non-Member Price
PDF $17.00 $21.00

Paper Abstract

In intelligent video surveillance, cameras record image sequences during day and night. Commonly, this demands different sensors. To achieve a better performance it is not unusual to combine them. We focus on the case that a long-wave infrared camera records continuously and in addition to this, another camera records in the visible spectral range during daytime and an intelligent algorithm supervises the picked up imagery. More accurate, our task is multispectral CNN-based object detection. At first glance, images originating from the visible spectral range differ between thermal infrared ones in the presence of color and distinct texture information on the one hand and in not containing information about thermal radiation that emits from objects on the other hand. Although color can provide valuable information for classification tasks, effects such as varying illumination and specialties of different sensors still represent significant problems. Anyway, obtaining sufficient and practical thermal infrared datasets for training a deep neural network poses still a challenge. That is the reason why training with the help of data from the visible spectral range could be advantageous, particularly if the data, which has to be evaluated contains both visible and infrared data. However, there is no clear evidence of how strongly variations in thermal radiation, shape, or color information influence classification accuracy. To gain deeper insight into how Convolutional Neural Networks make decisions and what they learn from different sensor input data, we investigate the suitability and robustness of different augmentation techniques. We use the publicly available large-scale multispectral ThermalWorld dataset consisting of images in the long-wave infrared and visible spectral range showing persons, vehicles, buildings, and pets and train for image classification a Convolutional Neural Network. The training data will be augmented with several modifications based on their different properties to find out which ones cause which impact and lead to the best classification performance.

Paper Details

Date Published: 7 October 2019
PDF: 8 pages
Proc. SPIE 11166, Counterterrorism, Crime Fighting, Forensics, and Surveillance Technologies III, 111660N (7 October 2019); doi: 10.1117/12.2532537
Show Author Affiliations
Vanessa Buhrmester, Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB (Germany)
Ann-Kristin Grosselfinger, Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB (Germany)
David Münch, Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB (Germany)
Michael Arens, Fraunhofer-Institut für Optronik, Systemtechnik und Bildauswertung IOSB (Germany)


Published in SPIE Proceedings Vol. 11166:
Counterterrorism, Crime Fighting, Forensics, and Surveillance Technologies III
Henri Bouma; Radhakrishna Prabhu; Robert James Stokes; Yitzhak Yitzhaky, Editor(s)

© SPIE. Terms of Use
Back to Top