Jump to content

H.261: Difference between revisions

From Wikipedia, the free encyclopedia
Content deleted Content added
m Reverted edits by 70.54.32.40 (talk) to last version by Jmkim dot com
Reim (talk | contribs)
mNo edit summary
Line 14: Line 14:
The basic processing unit of the design is called a [[macroblock]], and H.261 was the first standard in which the macroblock concept appeared. Each macroblock consists of a 16x16 array of [[Luminance (video)|luma]] samples and two corresponding 8x8 arrays of [[Chrominance|chroma]] samples, using [[Chroma subsampling|4:2:0 sampling]] and a [[YCbCr]] [[color space]].
The basic processing unit of the design is called a [[macroblock]], and H.261 was the first standard in which the macroblock concept appeared. Each macroblock consists of a 16x16 array of [[Luminance (video)|luma]] samples and two corresponding 8x8 arrays of [[Chrominance|chroma]] samples, using [[Chroma subsampling|4:2:0 sampling]] and a [[YCbCr]] [[color space]].


The inter-picture prediction reduces temporal redundancy, with motion vectors used to help the codec compensate for motion. Whilst only integer-valued motion vectors are supported in H.261, a blurring filter can be applied to the prediction signal — partially mitigating the lack of fractional-sample motion vector precision. Transform coding using an 8x8 [[discrete cosine transform]] (DCT) reduces the spatial redundancy. Scalar quantization is then applied to round the transform coefficients to the appropriate precision determined by a step size control parameter, and the quantized transform coefficients are zig-zag scanned and entropy coded (using a "run-level" variable-length code) to remove statistical redundancy.
The inter-picture prediction reduces temporal redundancy, with [[motion vector|motion vectors]] used to help the codec compensate for motion. Whilst only integer-valued motion vectors are supported in H.261, a blurring filter can be applied to the prediction signal — partially mitigating the lack of fractional-sample motion vector precision. Transform coding using an 8x8 [[discrete cosine transform]] (DCT) reduces the spatial redundancy. Scalar quantization is then applied to round the transform coefficients to the appropriate precision determined by a step size control parameter, and the quantized transform coefficients are zig-zag scanned and entropy coded (using a "run-level" variable-length code) to remove statistical redundancy.


The H.261 standard actually only specifies how to decode the video. Encoder designers were left free to design their own encoding algorithms, as long as their output was constrained properly to allow it to be decoded by any decoder made according to the standard. Encoders are also left free to perform any pre-processing they want to their input video, and decoders are allowed to perform any post-processing they want to their decoded video prior to display. One effective post-processing technique that became a key element of the best H.261-based systems is called [[deblocking]] filtering. This reduces the appearance of block-shaped artifacts caused by the block-based motion compensation and spatial transform parts of the design. Indeed, blocking artifacts are probably a familiar phenomenon to almost everyone who has watched digital video. Deblocking filtering has since become an integral part of the most recent standard, [[H.264]] (although even when using H.264, additional post-processing is still allowed and can enhance visual quality if performed well).
The H.261 standard actually only specifies how to decode the video. Encoder designers were left free to design their own encoding algorithms, as long as their output was constrained properly to allow it to be decoded by any decoder made according to the standard. Encoders are also left free to perform any pre-processing they want to their input video, and decoders are allowed to perform any post-processing they want to their decoded video prior to display. One effective post-processing technique that became a key element of the best H.261-based systems is called [[deblocking]] filtering. This reduces the appearance of block-shaped artifacts caused by the block-based motion compensation and spatial transform parts of the design. Indeed, blocking artifacts are probably a familiar phenomenon to almost everyone who has watched digital video. Deblocking filtering has since become an integral part of the most recent standard, [[H.264]] (although even when using H.264, additional post-processing is still allowed and can enhance visual quality if performed well).

Revision as of 16:19, 15 May 2010

H.261 is a 1990[citation needed] ITU-T video coding standard, ratified in November 1988.[1] Originally designed for transmission over ISDN lines on which data rates are multiples of 64 kbit/s. It is one member of the H.26x family of video coding standards in the domain of the ITU-T Video Coding Experts Group (VCEG). The coding algorithm was designed to be able to operate at video bit rates between 40 kbit/s and 2 Mbit/s. The standard supports two video frame sizes: CIF (352x288 luma with 176x144 chroma) and QCIF (176x144 with 88x72 chroma) using a 4:2:0 sampling scheme. It also has a backward-compatible trick for sending still picture graphics with 704x576 luma resolution and 352x288 chroma resolution (which was added in a later revision in 1993).

History

Whilst H.261 was preceded in 1984 by H.120 (which also underwent a revision in 1988 of some historic importance) as a digital video coding standard, H.261 was the first truly practical digital video coding standard (in terms of product support in significant quantities). In fact, all subsequent international video coding standards (MPEG-1 Part 2, H.262/MPEG-2 Part 2, H.263, MPEG-4 Part 2, and H.264/MPEG-4 Part 10) have been based closely on the H.261 design. Additionally, the methods used by the H.261 development committee to collaboratively develop the standard have remained the basic operating process for subsequent standardization work in the field (see S. Okubo, "Reference model methodology-A tool for the collaborative creation of video coding standards", Proceedings of the IEEE, vol. 83, no. 2, Feb. 1995, pp. 139–150). The coding algorithm uses a hybrid of motion compensated inter-picture prediction and spatial transform coding with scalar quantization, zig-zag scanning and entropy encoding.

H.261 design

The basic processing unit of the design is called a macroblock, and H.261 was the first standard in which the macroblock concept appeared. Each macroblock consists of a 16x16 array of luma samples and two corresponding 8x8 arrays of chroma samples, using 4:2:0 sampling and a YCbCr color space.

The inter-picture prediction reduces temporal redundancy, with motion vectors used to help the codec compensate for motion. Whilst only integer-valued motion vectors are supported in H.261, a blurring filter can be applied to the prediction signal — partially mitigating the lack of fractional-sample motion vector precision. Transform coding using an 8x8 discrete cosine transform (DCT) reduces the spatial redundancy. Scalar quantization is then applied to round the transform coefficients to the appropriate precision determined by a step size control parameter, and the quantized transform coefficients are zig-zag scanned and entropy coded (using a "run-level" variable-length code) to remove statistical redundancy.

The H.261 standard actually only specifies how to decode the video. Encoder designers were left free to design their own encoding algorithms, as long as their output was constrained properly to allow it to be decoded by any decoder made according to the standard. Encoders are also left free to perform any pre-processing they want to their input video, and decoders are allowed to perform any post-processing they want to their decoded video prior to display. One effective post-processing technique that became a key element of the best H.261-based systems is called deblocking filtering. This reduces the appearance of block-shaped artifacts caused by the block-based motion compensation and spatial transform parts of the design. Indeed, blocking artifacts are probably a familiar phenomenon to almost everyone who has watched digital video. Deblocking filtering has since become an integral part of the most recent standard, H.264 (although even when using H.264, additional post-processing is still allowed and can enhance visual quality if performed well).

Design refinements introduced in later standardization efforts have resulted in significant improvements in compression capability relative to the H.261 design. This has resulted in H.261 becoming essentially obsolete, although it is still used as a backward-compatibility mode in some video conferencing systems and for some types of internet video. However, H.261 remains a major historical milestone in the development of the field of video coding.

Software implementations

See also

References

  1. ^ "(Nokia position paper) Web Architecture and Codec Considerations for Audio-Visual Services" (PDF). H.261, which (in its first version) was ratified in November 1988.