Abstract:
Innovations in encoder-side decisions that use the results of hash-based block matching when setting parameters are presented. For example, some of the innovations relate to ways to select motion vector precision depending on the results of hash-based block matching. Other innovations relate to ways to selectively disable sample adaptive offset filtering depending on the results of hash-based block matching. Still other innovations relate to ways to select which reference pictures to retain in a reference picture set depending on the results of hash-based block matching.
Abstract:
Disclosed herein are representative embodiments of generating representative motion information that can be used during processing of a video frame. In one exemplary embodiment disclosed herein, a reference frame comprising a group of blocks is processed, and motion information for the group of blocks is compressed at least by buffering representative motion-vector information and representative reference-frame index information for the group of blocks. The representative reference-frame index information comprises reference-frame index information of a representative block of the group of blocks, and the representative reference-frame index information represents reference-frame index information for the group of blocks during processing of a current frame.
Abstract:
Innovations are provided for encoding and/or decoding video and/or image content using dictionary modes. For example, some innovations predict current pixel values from previous pixel values stored in a 1-D dictionary. Other innovations predict current pixel values from previous pixel values using a pseudo 2-D dictionary mode. Yet other innovations predict current pixel values from previous pixel values in a reference picture using an inter pseudo 2-D dictionary mode. Pixel values can be predicted from previous pixel values (e.g., stored in a dictionary) that are identified by an offset and a length. Yet other innovations encode pixel values using hash matching of pixel values.
Abstract:
Embodiments for reducing redundancy in video coding are disclosed. In accordance with at least one embodiment, video content is represented as a tree structure in which the nodes of the tree structure are associated with attributes of the video content. An encoder analyzes the nodes to determine whether there is redundancy associated with a last node of a divided node. In another embodiment, bidirectional prediction video coding techniques are applied to the video content. The encoder analyzes two lists of reference frames to determine whether a third mode of bidirectional prediction is redundant to a second mode of bidirectional prediction. The efficiency of video coding is improved by refraining from sending data symbols to the decoder in the event that the video coding techniques contain redundancy.
Abstract:
Disclosed herein are representative embodiments for performing entropy coding or decoding using a counter-based scheme. In one exemplary embodiment disclosed herein, a first codeword is received from compressed digital media data. The first codeword is decoded into a first digital media data value by referencing a codeword table that associates the first codeword with the first digital media data value and a second codeword with a second digital media data value. A counter for counting occurrences of the first digital media data value is incremented. The value of the first counter is compared with the value of a second counter that counts occurrences of a second digital media data value. If the value of the first counter and the value of the second counter are equal (or greater than or equal), the codeword table is updated to swap codewords between the first and second digital media values.
Abstract:
Techniques for implementing video encoding enhancements may increase video encoding efficiency. One of the techniques may involve the use of an exception value and a maximum index value in inter-prediction coding lookup table to enhance the combined coding of an inter-prediction direction and a reference frame index for a coding unit. Another of the techniques may include adaptively ordering the encoding of a split flag value and a skip flag value for a coding unit. An additional technique may include providing a uniform approach to adaptively combined code the coded block flag (CBF) values for transform or coding units of images that are used with the Residue Quad-tree (RQT) information to represent residue information.
Abstract:
Various innovations in media encoding are presented herein. In particular, the innovations can reduce the computational complexity of encoding by selectively skipping certain evaluation stages during encoding. For example, based on analysis of decisions made earlier in encoding or based on analysis of media to be encoded, an encoder can selectively skip evaluation of certain coding tools (such as residual coding or rate-distortion-optimized quantization), skip evaluation of certain values for parameters or settings (such as candidate unit sizes or transform sizes, or candidate partition patterns for motion compensation), and/or skip evaluation of certain coding modes (such as frequency transform skip mode) that are not expected to improve rate-distortion performance during encoding.
Abstract:
Innovations in the area of encoding or decoding of blocks using intra block copy (“BC”) prediction improve coding efficiency for intra-BC-predicted blocks in many situations. For example, some of the innovations relate to block flipping in which an intra-BC-predicted block is flipped relative to a reference region, which can be indicated by a block vector (“BV”) value. Other innovations relate to signaling of a skip mode in which a current intra-BC-predicted block uses a signaled BV differential but lacks residual data.
Abstract:
Innovations in intra block copy (“BC”) prediction mode facilitate intra BC prediction that is more effective in terms of rate-distortion performance and/or computational efficiency of encoding and decoding. For example, some of the innovations provide ways to select block vector (“BV”) values more effectively. Other innovations provide ways to encode/decode BV values more efficiently. Still other innovations address how to perform in-loop deblock filtering when a block has intra BC prediction mode, or address how to perform intra BC prediction when constrained intra prediction is enabled.
Abstract:
Techniques for coding and deriving (e.g., determining) one or more coded-block-flags associated with video content are described herein. A coded-block-flag of a last node may be determined when coded-block-flags of preceding nodes are determined to be a particular value and when a predetermined condition is satisfied. In some instances, the predetermined condition may be satisfied when log2(size of current transform unit) is less than log2(size of maximum transform unit) or log2(size of current coding unit) is less than or equal to log2(size of maximum transform unit)+1. The preceding nodes may be nodes that precede the last node on a particular level in a residual tree.