Abstract:
Techniques are described to improved video intra prediction using position-dependent prediction combination in video coding. In High Efficiency Video Encoding a set of 35 linear predictors are used for doing intra coding and prediction can be computed from either a nonfiltered or a filtered set of neighboring “reference” pixels, depending on the selected predictor mode and block size. Techniques of this disclosure may use a weighted combination of both the nonfiltered and filtered set of reference pixels to achieve better compression via improved prediction and therefore small residual, enable effective parallel computation of all sets of prediction values, and maintain low complexity via applying filtering only to a set of reference pixels and not to predicted values themselves.
Abstract:
A device for decoding video data includes a memory configured to store the video data; and one or more processors configured to decode syntax information that indicates a selected intra prediction mode for the block of video data from among a plurality of intra prediction modes. The one or more processors apply an N-tap intra interpolation filter to neighboring reconstructed samples of the block of video data according to the selected intra prediction mode, wherein N is greater than 2. The one or more processors reconstruct the block of video data based on the filtered neighboring reconstructed samples according to the selected intra prediction mode.
Abstract:
Techniques and systems are provided for coding video data. For example, a method of coding video data includes determining one or more illumination compensation parameters for a current block and coding the current block as part of an encoded bitstream using the one or more illumination compensation parameters. In some cases, the method can include determining one or more spatially neighboring samples for the current block and deriving the one or more illumination compensation parameters for the current block based on at least one of the one or more spatially neighboring samples. The method can further include signaling, individually, for the current block, an illumination compensation status in the encoded bitstream. The method can further include signaling at least one of the one or more illumination compensation parameters for the current block in the encoded bitstream.
Abstract:
An apparatus for coding video information according to certain aspects includes a memory unit and a processor in communication with the memory unit. The memory unit stores difference video information associated with a difference video layer of pixel information derived from a difference between an enhancement layer and a corresponding base layer of the video information. The processor determines an enhancement layer weight and a base layer weight, and determines a value of a current video unit based on the difference video layer, a value of a video unit in the enhancement layer weighted by the enhancement layer weight, and a value of a video unit in the base layer weighted by the base layer weight.
Abstract:
A video coding device includes a memory configured to store video data and processor(s) configured to process at least a portion of the stored video data. The processor(s) are configured to identify a coefficient group (CG) that includes a current transform coefficient of the video data, the CG representing a subset of transform coefficients within a transform unit. The processor(s) are further configured to determine a size of the CG based on a combination of a transform size and one or both of (i) a coding mode associated with the transform unit, or (ii) a transform matrix associated with the transform unit.
Abstract:
Context adaptive binary arithmetic coding (CABAC) techniques are generally described. Aspects of the techniques are generally directed to inheritance-based context initialization. An example video coding device includes a memory configured to store video data, and one or more processors. The processor(s) are configured to initialize context information for a current slice of a current picture by inheriting context information of a previously-coded block of a previously-coded picture of the stored video data as initialized context information for the current slice of the current picture. The processor(s) are further configured to code data of the current slice using the initialized context information.
Abstract:
An example method of entropy coding video data includes determining a window size of a plurality of window sizes for a context of a plurality of contexts used in a context-adaptive coding process to entropy code a value for a syntax element of the video data; entropy coding, based on a probability state of the context, a bin of the value for the syntax element; updating a probability state of the context based on the window size and the coded bin. The example method also includes entropy coding a next bin with the same context based on the updated probability state of the context.
Abstract:
An example method of entropy coding video data includes obtaining a pre-defined initialization value for a context of a plurality of contexts used in a context-adaptive entropy coding process to entropy code a value for a syntax element in a slice of the video data, wherein the pre-defined initialization value is stored with N-bit precision; determining, using a look-up table and based on the pre-defined initialization value, an initial probability state of the context for the slice of the video data, wherein a number of possible probability states for the context is greater than two raised to the power of N; and entropy coding, based on the initial probability state of the context, a bin of the value for the syntax element.
Abstract:
In one example, a device for coding video data includes a video coder configured to code data indicating whether tile boundaries of different layers of video data are aligned and whether inter-layer prediction is allowed along or across tile boundaries of enhancement layer blocks, code an enhancement layer block in an enhancement layer tile of the video data without using inter-layer prediction from a collocated base layer block for which inter-layer filtering or reference layer filtering across tile boundaries in a reference layer picture in an access unit including both the enhancement layer tile and the base layer block is enabled, and code the collocated base layer block.
Abstract:
As one example, techniques for decoding video data include receiving a bitstream that includes one or more pictures of a coded video sequence (CVS), decoding a first picture according to a decoding order, wherein the first picture is a random access point (RAP) picture that is not an instantaneous decoding refresh (IDR) picture, and decoding at least one other picture following the first picture according to the decoding order based on the decoded first picture. As another example, techniques for encoding video data include generating a bitstream that includes one or more pictures of a CVS, wherein a first picture according to the decoding order is a RAP picture that is not an IDR picture, and avoiding including at least one other picture, other than the first picture, that corresponds to a leading picture associated with the first picture, in the bitstream.