Share Email Print

Optical Engineering

Story-related caption detection and localization in news video
Author(s): Chien-Cheng Lee; Cheng-Yuan Shih; Hao-Ming Huang
Format Member Price Non-Member Price
PDF $20.00 $25.00
cover GOOD NEWS! Your organization subscribes to the SPIE Digital Library. You may be able to download this paper for free. Check Access

Paper Abstract

We propose a method to detect and localize story-related subject captions in news video. Most caption detection and localization algorithms attempt to detect as many captions as possible; however, a news picture may contain many types of captions that are unrelated to the story. To facilitate fast and accurate access to news video content, a method for detecting and localizing the story-related caption is necessary. This paper addresses two problems in texture-based caption detection and localization: the time-consuming computation of features, and the clutter of caption detection results. We address these problems by first identifying the subject caption region based on the frequency of text occurrence. Then, we detect the subject caption frame as it first appears onscreen. Finally, the texture-based caption localization procedure is performed on the subject caption region in subject captions' beginning frames. Using this method decreases the computation time significantly. Additionally, the unrelated types of text are also filtered out, and only the subject caption is detected and localized. Experimental results show that the proposed method can quickly and robustly detect subject captions from news video.

Paper Details

Date Published: 1 March 2009
PDF: 13 pages
Opt. Eng. 48(3) 037005 doi: 10.1117/1.3103126
Published in: Optical Engineering Volume 48, Issue 3
Show Author Affiliations
Chien-Cheng Lee, Yuan Ze Univ. (Taiwan)
Cheng-Yuan Shih, Yuan Ze Univ. (Taiwan)
Hao-Ming Huang, Yuan Ze Univ. (Taiwan)

© SPIE. Terms of Use
Back to Top