Abstract:
A video block syntax element indicates whether all of the partitions of a video block are predicted based on a same reference list and no greater than quarter-pixel accuracy is used. If the video block syntax element is set, partition-level signaling of the reference lists is avoided. If the video block syntax element is not set, partition-level signaling of the reference lists occurs. If the video block syntax element is set, partition-level syntax elements may be used for each of the partitions of the video block, wherein the partition-level syntax elements each identify one of the reference lists and motion vector accuracy for a given one of the partitions.
Abstract:
This disclosure describes techniques for coding an enhancement layer in a scalable video coding (SVC) scheme. The techniques involve run-length coding of significant coefficients and refinement coefficients of the enhancement layer. Rather than performing two different run-length passes to separately code the significant coefficients and refinement coefficients, the techniques of this disclosure perform run-length coding of the significant coefficients and refinement coefficients together. Therefore, run values of the run-length coding codes the significant coefficients with the refinement coefficients. Additional techniques are also described, which can eliminate the need to send sign information for some of the refinement coefficients. Instead, this sign information for some of the refinement coefficients may be derived at the decoder based on the sign values of corresponding coefficients of previously encoded layers of the SVC scheme, which can further improve compression efficiency.
Abstract:
Techniques are described for encoding and decoding digital video data using macroblocks that are larger than the macroblocks prescribed by conventional video encoding and decoding standards. For example, the techniques include encoding and decoding a video stream using macroblocks comprising greater than 16×16 pixels. In one example, an apparatus includes a video encoder configured to encode a coded unit comprising a plurality of video blocks, wherein at least one of the plurality of video blocks comprises a size of more than 16×16 pixels and to generate syntax information for the coded unit that includes a maximum size value, wherein the maximum size value indicates a size of a largest one of the plurality of video blocks in the coded unit. The syntax information may also include a minimum size value. In this manner, the encoder may indicate to a decoder the proper syntax decoder to apply to the coded unit.
Abstract:
This disclosure describes techniques for scanning coefficients of video blocks. In particular, the techniques of this disclosure adapt a scan order used to scan a two-dimensional block of coefficients into a one-dimensional coefficient vector based on statistics associated with one or more previously coded blocks. For example, statistics that indicate the likelihood that a given coefficient value in each position of a two-dimensional block is zero or non-zero may be collected for one or more previously coded blocks. At some point, an adjustment to the scan order can be made in order to better ensure that non-zero coefficients are grouped together near the front of the one-dimensional coefficient vector, which can improve the effectiveness of entropy coding. The collection of statistics and adjustment of scan order may be made separately for each possible prediction mode.
Abstract:
In one example, a video coder is configured to code information indicative of whether view synthesis prediction is enabled for video data. When the information indicates that view synthesis prediction is enabled for the video data, the video coder may generate a view synthesis picture using the video data and code at least a portion of a current picture relative to the view synthesis picture. The at least portion of the current picture may comprise, for example, a block (e.g., a PU, a CU, a macroblock, or a partition of a macroblock), a slice, a tile, a wavefront, or the entirety of the current picture. On the other hand, when the information indicates that view synthesis prediction is not enabled for the video data, the video coder may code the current picture using at least one of intra-prediction, temporal inter-prediction, and inter-view prediction without reference to any view synthesis pictures.
Abstract:
This disclosure describes techniques for entropy coding of video blocks, and proposes a syntax element that may promote coding efficiency. The syntax element may identify a number of non-zero value sub-blocks within a video block, wherein the non-zero value sub-blocks comprise sub-blocks within the video block that include at least one non-zero coefficient. A method of coding a video block may comprise coding the syntax element, generating the non-zero value sub-blocks of the video block, and entropy coding the non-zero value sub-blocks.
Abstract:
This disclosure describes video coding techniques that are executed in a bitdepth-based scalable video coding (SVC) scheme. Base layer video data and enhancement layer video data may be defined at different bitdepths. The techniques of this disclosure concern the coding of enhancement layer video blocks using a unique combination of inter-layer predictive coding modes after determining that inter-layer prediction should be used. In particular, two or more different inter-layer predictive coding modes may be utilized block corresponding to inter base layer blocks, and one of these inter-layer predictive coding modes may also be utilized for blocks corresponding to intra base layer blocks. Syntax information can be generated at the encoder and communicated with the bitstream, so that a decoder can use the proper inter-layer predictive coding mode in the decoding.
Abstract:
This disclosure describes techniques for coding of header information of video blocks. In particular, the techniques of this disclosure select one of a plurality of prediction modes for use in generating a prediction block of a video block of a coding unit, the plurality of prediction modes including unidirectional prediction modes and multi-directional prediction modes that combine at least two unidirectional prediction modes. An encoding device encodes the prediction mode of the current video block based on prediction modes of one or more previously encoded video blocks of the coding unit. Likewise, a decoding unit receives encoded video data of a video block of a coding unit and decodes the encoded video data to identify one of a plurality of prediction modes for use in generating a prediction block of the video block based on prediction modes of one or more previously decoded video blocks of the coding unit.
Abstract:
This disclosure describes techniques for coding transform coefficients for a block of video data. For example, according to one embodiment, a video encoder determines an lrg1Pos value associated with the transform coefficient based on the noTr1 value and a position k of the transform in the scan order of the block of video data based on using at least one table that defines an lrg1Pos value for more than one potential noTr1 value for the scan order of the block of video data. In one embodiment, the video decoder uses the determined lrg1Pos value associated with the transform coefficient to perform a structured mapping to determine a code number cn based on a determined value for the level_ID syntax element and a determined value for the run syntax element.
Abstract:
Techniques for transforming the prediction error of intra-coded blocks using mode-dependent transform functions. In an embodiment, an encoder selects a set of transform functions to represent prediction error based on the spatial mode used for prediction. In an alternative embodiment, a decoder reconstructs an image block by using the signaled spatial mode to derive the corresponding set of transform functions. No additional signaling between encoder and decoder is required as compared to prior art implementations.