Abstract:
Techniques and tools for performing fading compensation in video processing applications are described. For example, during encoding, a video encoder performs fading compensation using fading parameters comprising a scaling parameter and a shifting parameter on one or more reference images. During decoding, a video decoder performs corresponding fading compensation on the one or more reference images.
Abstract:
Techniques and tools are described for compensating for rounding when estimating sample-domain distortion in the transform domain. For example, a video encoder estimates pixel-domain distortion in the transform domain for a block of transform coefficients after compensating for rounding in the DC coefficient of the block. In this way, the video encoder improves the accuracy of pixel-domain distortion estimation but retains the computational advantages of performing the estimation in the transform domain. Rounding compensation includes, for example, looking up an index (from a de-quantized transform coefficient) in a rounding offset table to determine a rounding offset, then adjusting the coefficient by the offset. Other techniques and tools described herein are directed to creating rounding offset tables and encoders that make encoding decisions after considering rounding effects that occur after an inverse frequency transform on de-quantized transform coefficient values.
Abstract:
Techniques and tools for switching distortion metrics during motion estimation are described. For example, a video encoder determines a distortion metric selection criterion for motion estimation. The criterion can be based on initial results of the motion estimation. To evaluate the criterion, the encoder can compare the criterion to a threshold that depends on a current quantization parameter. The encoder selects between multiple available distortion metrics, which can include a sample-domain distortion metric (e.g., SAD) and a transform-domain distortion metric (e.g., SAHD). The encoder uses the selected distortion metric in the motion estimation. Selectively switching between SAD and SAHD provides rate-distortion performance superior to using only SAD or only SAHD. Moreover, due to the lower complexity of SAD, the computational complexity of motion estimation with SAD-SAHD switching is typically less than motion estimation that always uses SAHD.
Abstract:
Techniques and tools are described for compensating for rounding when estimating sample-domain distortion in the transform domain. For example, a video encoder estimates pixel-domain distortion in the transform domain for a block of transform coefficients after compensating for rounding in the DC coefficient of the block. In this way, the video encoder improves the accuracy of pixel-domain distortion estimation but retains the computational advantages of performing the estimation in the transform domain. Rounding compensation includes, for example, looking up an index (from a de-quantized transform coefficient) in a rounding offset table to determine a rounding offset, then adjusting the coefficient by the offset. Other techniques and tools described herein are directed to creating rounding offset tables and encoders that make encoding decisions after considering rounding effects that occur after an inverse frequency transform on de-quantized transform coefficient values.
Abstract:
Rules for the signaling and interpretation of chroma position are described. One rule, called the short rule, defines fifteen discrete chroma centering positions and corresponding four-bit syntax element. Another rule, called the extended rule, defines 81 discrete chroma centering positions and corresponding seven-bit syntax elements. A described method includes receiving digital media data at a digital media encoder, determining chroma position information for the received digital media data, and representing the chroma position information with one or more syntax elements in an encoded bitstream. The one or more syntax elements are operable to communicate the chroma position information to a digital media decoder. The chroma position information facilitates an image rotation or flip.
Abstract:
In certain embodiments, overlap operators are applied during encoding and/or decoding of digital media, where the overlap operators have reduced DC gain mismatch and/or DC leakage between interior overlap operators and overlap operators at the edge and/or corner. In other embodiments, information indicating a selected tile boundary option for overlap processing can be encoded and/or decoded. The selected tile boundary option indicates one of a hard tile boundary option and a soft tile boundary option for processing with overlap operators. Overlap transform processing can then be applied based at least in part on the selected tile boundary option.
Abstract:
Techniques and tools for conversion operations between modules in a scalable video encoding tool or scalable video decoding tool are described. For example, given reconstructed base layer video in a low resolution format (e.g., 4:2:0 video with 8 bits per sample) an encoding tool and decoding tool adaptively filter the reconstructed base layer video and upsample its sample values to a higher sample depth (e.g., 10 bits per sample). The tools also adaptively scale chroma samples to a higher chroma sampling rate (e.g., 4:2:2). The adaptive filtering and chroma scaling help reduce energy in inter-layer residual video by making the reconstructed base layer video closer to input video, which typically makes compression of the inter-layer residual video more efficient. The encoding tool also remaps sample values of the inter-layer residual video to adjust dynamic range before encoding, and the decoding tool performs inverse remapping after decoding.
Abstract:
Architecture for enhancing the compression (e.g., luma, chroma) of a video signal and improving the perceptual quality of the video compression schemes. The architecture operates to reshape the normal multimodal energy distribution of the input video signal to a new energy distribution. In the context of luma, the algorithm maps the black and white (or contrast) information of a picture to a new energy distribution. For example, the contrast can be enhanced in the middle range of the luma spectrum, thereby improving the contrast between a light foreground object and a dark background. At the same time, the algorithm reduces the bit-rate requirements at a particular quantization step size. The algorithm can be utilized also in post-processing to improve the quality of decoded video.
Abstract:
Techniques and tools for switching distortion metrics during motion estimation are described. For example, a video encoder determines a distortion metric selection criterion for motion estimation. The criterion can be based on initial results of the motion estimation. To evaluate the criterion, the encoder can compare the criterion to a threshold that depends on a current quantization parameter. The encoder selects between multiple available distortion metrics, which can include a sample-domain distortion metric (e.g., SAD) and a transform-domain distortion metric (e.g., SAHD). The encoder uses the selected distortion metric in the motion estimation. Selectively switching between SAD and SAHD provides rate-distortion performance superior to using only SAD or only SAHD. Moreover, due to the lower complexity of SAD, the computational complexity of motion estimation with SAD-SAHD switching is typically less than motion estimation that always uses SAHD.
Abstract:
A video decoder receives an entry point key frame comprising first and second interlaced video fields and decodes a first syntax element comprising information (e.g., frame coding mode) for the entry point key frame at a first syntax level (e.g., frame level) in a bitstream. The first interlaced video field is a predicted field, and the second interlaced video field is an intra-coded field. The information for the entry point key frame can be a frame coding mode (e.g., field interlace) for the entry point key frame. The decoder can decode a second syntax element at the first syntax level comprising second information (e.g., field type for each of the first and second interlaced video fields) for the entry point key frame.