Abstract:
In one example, a device for coding video data includes a video coder configured to code data indicating whether tile boundaries of different layers of video data are aligned and whether inter-layer prediction is allowed along or across tile boundaries of enhancement layer blocks, code an enhancement layer block in an enhancement layer tile of the video data without using inter-layer prediction from a collocated base layer block for which inter-layer filtering or reference layer filtering across tile boundaries in a reference layer picture in an access unit including both the enhancement layer tile and the base layer block is enabled, and code the collocated base layer block.
Abstract:
In one example, a device for coding video data includes a video coder configured to code data representative of whether a tile of an enhancement layer picture can be predicted using inter-layer prediction, and predict data of the tile using inter-layer prediction only when the data indicates that the tile can be predicted using inter-layer prediction.
Abstract:
This disclosure presents techniques for sample adaptive offset (SAO) filtering that focus on filtering the changed region of a picture or layer (e.g., the high frequency component of video blocks of a picture or layer). In general, the techniques of this disclosure may be referred to as High-Frequency-Pass SAO (HSAO) filtering. In one example, a method for coding video data comprises obtaining a decoded block of video data, separating the decoded block of video data into a low-frequency component and a high-frequency component, applying an SAO filter to the high-frequency component to generate a SAO filtered high-frequency component, and combining the SAO filtered high-frequency component and the low-frequency component to generate a filtered decoded block of video data.
Abstract:
Methods and systems for video image coding are provided. Sets of filters may be selected and applied to video information at least partially based on the type of inter layer prediction implemented in coding the video information. Different filters, or filter sets, may be used for inter layer intra prediction, difference domain intra prediction, and/or difference domain inter prediction. Filter selection information may be embedded in the video bit stream.
Abstract:
A video coding process that includes defining a context derivation neighborhood for one of a plurality of transform coefficients based on a transform coefficient scan order. The process also includes determining a context for the one of the plurality of transform coefficients based on the context derivation neighborhood. The process also includes coding the one of the plurality of transform coefficients based on the determined context.
Abstract:
An example video encoder is configured to receive an indication of merge mode coding of a block within a parallel motion estimation region (PMER), generate a merge mode candidate list comprising one or more spatial neighbor motion vector (MV) candidates and one or more temporal motion vector prediction (TMVP) candidates, wherein motion information of at least one of the spatial neighbor MV candidates is known to be unavailable during coding of the block at an encoder, determine an index value identifying, within the merge mode candidate list, one of the TMVP candidates or the spatial neighbor MV candidates for which motion information is available during coding of the particular block, and merge mode code the block using the identified MV candidate.
Abstract:
In general, techniques are described for performing motion vector prediction for video coding. A video coding device comprising a processor may perform the techniques. The processor may be configured to determine a plurality of candidate motion vectors for a current block of the video data so as to perform the motion vector prediction process and scale one or more of the plurality of candidate motion vectors determined for the current block of the video data to generate one or more scaled candidate motion vectors. The processor may then be configured to modify the scaled candidate motion vectors to be within a specified range.
Abstract:
A device for video decoding can be configured to obtain, from a syntax structure in a bitstream comprising an encoded representation of the video data, a syntax element indicating whether 6-parameter affine prediction is enabled for blocks corresponding to the syntax structure, wherein the blocks corresponding to the syntax structure comprise a first block; based on the syntax element indicating that the 6-parameter affine prediction is enabled for the blocks corresponding to the syntax structure, use the 6-parameter affine prediction to generate a predictive block for the first block; and use the predictive block and residual data to reconstruct the first block.
Abstract:
A device for encoding video data includes a memory configured to store video data, and a video encoder implemented in circuitry and configured to encode a future picture of the video data having a first display order position, the future picture being included in an intra period (IP) of the video data, the IP comprising a plurality of groups of pictures (GOPs), and after encoding the future picture, encode a picture of an ordinal first GOP of the plurality of GOPs using the future picture as a reference picture, each picture of the ordinal first GOP having display order positions earlier than the first display order position. Encoding the future picture in this manner may result in encoding performance improvements with minimal increases in encoding and decoding complexity.
Abstract:
An example device for filtering decoded video data includes a memory configured to store video data; and one or more processors implemented in circuitry and configured to: decode a picture of video data; code a value for a syntax element representing a neural network model to be used to filter a portion of the decoded picture, the value representing an index into a set of pre-defined neural network models, the index corresponding to the neural network model in the set of pre-defined neural network models; and filter the portion of the decoded picture using the neural network model corresponding to the index.