Share Email Print
cover

Proceedings Paper

Extraction of moving objects for content-based video coding
Author(s): Thomas Meier; King N. Ngan
Format Member Price Non-Member Price
PDF $14.40 $18.00
cover GOOD NEWS! Your organization subscribes to the SPIE Digital Library. You may be able to download this paper for free. Check Access

Paper Abstract

This paper considers video object plane (VOP) segmentation for the content-based video coding standard MPEG-4. To provide multimedia applications with new functionalities, such as content-based interactivity and scalability, the new video coding standard MPEG-4 relies on a content-based representation. To take advantage of these functionalities, a prior decomposition of sequences into semantically meaningful, physical objects is required. We formulate this problem as one of separating foreground objects from the background based on motion information. For the object of interest, a two- dimensional binary model is derived and tracked throughout the sequence. The model points consist of edge pixels detected by the Canny operator. To accommodate rotation and changes in shape of the tracked object, the model is updated every frame. These binary models then guide the actual VOP extraction. Due to the excellent edge localization properties of the Canny operator, the resulting VOP contours are very accurate. Both the model initialization and update stage exploit motion information. The main assumption underlying our approach is the existence of a dominant global motion that can be assigned to the background. Areas that do not follow this background motion indicate the presence of independently moving physical objects. Two methods to identify such objects are presented. The first one employs a morphological motion filter with a new filtering criterion that measures the deviation of the locally estimated optical flow from the corresponding global motion. The second method computes a change detection mask by taking the difference between consecutive frames. The first version is more suitable for sequences involving little motion, whereas the second version is stronger at dealing with fast moving objects.

Paper Details

Date Published: 28 December 1998
PDF: 12 pages
Proc. SPIE 3653, Visual Communications and Image Processing '99, (28 December 1998); doi: 10.1117/12.334624
Show Author Affiliations
Thomas Meier, Univ. of Western Australia (United States)
King N. Ngan, Univ. of Western Australia (Australia)


Published in SPIE Proceedings Vol. 3653:
Visual Communications and Image Processing '99
Kiyoharu Aizawa; Robert L. Stevenson; Ya-Qin Zhang, Editor(s)

© SPIE. Terms of Use
Back to Top