Share Email Print
cover

Proceedings Paper

Pair-match Huffman transcoding to achieve a highly parallel variable-length decoder with two-word bit stream segmentation
Author(s): Michael Bakhmutsky
Format Member Price Non-Member Price
PDF $14.40 $18.00

Paper Abstract

An MPEG-2 Main Profile, high level compliant HDTV video decoder requires a variable length decoder (VLD) that can decode macroblocks at rates exceeding 100 million code words per second. The implementation of a high-performance VLD for such an application presents a major challenge in architecture design. The capability of the VLD to process macroblocks in real-time can reduce system memory requirements and simplify decoder architectures. It is more desirable to conceive of a 'one-piece' VLD capable of operating with minimal logic and memory resources and in real-time. Parallel partitioning of the video decoder on the VLD level increases the overall complexity and memory utilization. The two-word bit stream segmentation method developed by Philips Research -- USA achieves high performance without the expense of high hardware complexity or the addition of extra system memory. In this respect, this VLD implementation is very suitable for consumer digital HDTV video decoders. However, the performance guarantee of this architecture is associated with carefully specified statistical tradeoffs. The described method of pair-mach Huffman transcoding provides the VLD performance guarantee on a macroblock level without any statistical tradeoffs. Applied to the main body of the bit stream, this method produces excellent performance results for both consumer and professional MPEG profiles.

Paper Details

Date Published: 17 January 1997
PDF: 19 pages
Proc. SPIE 3021, Multimedia Hardware Architectures 1997, (17 January 1997); doi: 10.1117/12.263517
Show Author Affiliations
Michael Bakhmutsky, Philips Research (United States)


Published in SPIE Proceedings Vol. 3021:
Multimedia Hardware Architectures 1997
Sethuraman Panchanathan; Frans Sijstermans, Editor(s)

© SPIE. Terms of Use
Back to Top