Share Email Print
cover

Proceedings Paper

Prediction of visual saliency in video with deep CNNs
Author(s): Souad Chaabouni; Jenny Benois-Pineau; Ofer Hadar
Format Member Price Non-Member Price
PDF $14.40 $18.00

Paper Abstract

Prediction of visual saliency in images and video is a highly researched topic. Target applications include Quality assessment of multimedia services in mobile context, video compression techniques, recognition of objects in video streams, etc. In the framework of mobile and egocentric perspectives, visual saliency models cannot be founded only on bottom-up features, as suggested by feature integration theory. The central bias hypothesis, is not respected neither. In this case, the top-down component of human visual attention becomes prevalent. Visual saliency can be predicted on the basis of seen data. Deep Convolutional Neural Networks (CNN) have proven to be a powerful tool for prediction of salient areas in stills. In our work we also focus on sensitivity of human visual system to residual motion in a video. A Deep CNN architecture is designed, where we incorporate input primary maps as color values of pixels and magnitude of local residual motion. Complementary contrast maps allow for a slight increase of accuracy compared to the use of color and residual motion only. The experiments show that the choice of the input features for the Deep CNN depends on visual task:for th eintersts in dynamic content, the 4K model with residual motion is more efficient, and for object recognition in egocentric video the pure spatial input is more appropriate.

Paper Details

Date Published: 28 September 2016
PDF: 14 pages
Proc. SPIE 9971, Applications of Digital Image Processing XXXIX, 99711Q (28 September 2016); doi: 10.1117/12.2238956
Show Author Affiliations
Souad Chaabouni, LaBRI (France)
REGIM-Lab (Tunisia)
Jenny Benois-Pineau, LaBRI (France)
Ofer Hadar, Ben-Gurion Univ. of the Negev (Israel)


Published in SPIE Proceedings Vol. 9971:
Applications of Digital Image Processing XXXIX
Andrew G. Tescher, Editor(s)

© SPIE. Terms of Use
Back to Top