PolyU Institutional Repository >
Electronic and Information Engineering >
EIE Conference Papers & Presentations >
Please use this identifier to cite or link to this item:
|Title: ||New pixel-DCT domain coding technique for object based and frame based prediction error|
|Authors: ||Hui, Ko-cheung|
|Subjects: ||Discrete cosine transform|
DCT based coding
|Issue Date: ||2005 |
|Citation: ||2005 IEEE International Conference on Acoustics, Speech, and Signal Processing : proceedings : March 18-23, 2005, Philadelphia, Pennsylvania, USA, p. II377-II380.|
|Abstract: ||Discrete Cosine Transform (DCT) is widely used in modern video compression standards, such as the ITU-T H.263 and the ISO MPEG-4, to achieve high compression efficiency. A major merit of the DCT is its capability in
high energy compaction for natural images. However, the motion prediction error frame is not a natural image but synthetically generated by the process of motion compensation. This process degrades the energy compaction efficiency of the DCT. We study the spatial distribution of the prediction errors resulting from either
the full-search motion estimation or other fast search algorithms in order to improve the coding efficiency of the DCT. Subsequently, a Mixed Spatial-DCT-based Coding Scheme is proposed for coding the prediction errors in this paper. Our experimental results show that this coding
scheme can successfully improve the compression performance of the traditional DCT-based video coder with block based motion compensation for arbitrary shaped video objects and, video sequences which contain moderate to high motion activities.|
|Rights: ||© 2005 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.|
This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders. All persons copying this information are expected to adhere to the terms and constraints invoked by each author's copyright. In most cases, these works may not be reposted without the explicit permission of the copyright holder.
|Type: ||Conference Paper|
|Appears in Collections:||EIE Conference Papers & Presentations|
All items in the PolyU Institutional Repository are protected by copyright, with all rights reserved, unless otherwise indicated.
No item in the PolyU IR may be reproduced for commercial or resale purposes.