Abstract:
Encoding and decoding using spatial proximity context entropy coding may include identifying a plurality of transform coefficients for a current block of a current frame of a video stream. The plurality of transform coefficients may be ordered based on a scan order. A current transform coefficient may be identified from the plurality of transform coefficients. A plurality of context coefficients may be identified from the plurality of transform coefficients. Each context coefficient may be spatially proximate to the current transform coefficient and may be available for entropy coding the current transform coefficient. An entropy coding probability for the current transform coefficient may be identified based on the scan order and the plurality of context coefficients. The current transform coefficient may be entropy coded based on the entropy coding probability. The entropy coded current transform coefficient may be included in an output bitstream, which may be stored or transmitted.
Abstract:
A system and method provide a video description length (VDL) guided constant quality video encoding strategy with bitrate constraint and a video coding system for optimizing encoding bitrate, distortion and complexity of an input video. The method obtains an overall VDL, temporal VDL and spatial VDL of the input video and compares the overall VDL, temporal VDL and spatial VDL of the input video with a reference VDL, temporal VDL and spatial VDL. Based on the comparison, the method adjusts the encoding bitrate, the overall encoding complexity, temporal encoding complexity and spatial encoding complexity of the input video and encodes the input video with the adjusted encoding bitrate, overall encoding complexity, temporal encoding complexity and spatial encoding complexity of the input video.
Abstract:
A two-dimensional image to be converted to a first three-dimensional image may be received. A second three-dimensional image that is visually similar to the two-dimensional image that is to be converted may be identified. A feature-to-depth mapping function may be computed for the first three-dimensional image by using an approximate depth map of the second three-dimensional image that is visually similar to the two-dimensional image that is to be converted. The feature-to-depth mapping function may be applied to a plurality of pixels of the two-dimensional image to determine a depth value for the plurality of pixels of the two-dimensional image. The first three-dimensional image may be generated based on the depth values for the plurality of pixels of the two-dimensional image.
Abstract:
Super-transform coding may include identifying a plurality of sub-blocks for prediction coding a current block, determining whether to encode the current block using a super-transform, and super-prediction coding the current block. Super-prediction coding may include generating a super-prediction block for the current block by generating a prediction block for each unpartitioned sub-block of the current block, generating a super-prediction block for each partitioned sub-block of the current block by super-prediction coding the sub-block, and including the prediction blocks and super-prediction blocks for the sub-blocks in a super-prediction block for the current block. Including the prediction blocks and super-prediction blocks for the sub-blocks in a super-prediction block for the current block may include filtering at least a portion of each prediction block and each super-prediction block based on a spatially adjacent prediction block. Super-transform coding may include transforming the super-prediction block for the current block using a corresponding super-transform.
Abstract:
Coding a current block using multi-level compound predictor is disclosed. A method includes generating a first compound predictor by combining at least two predictor blocks for the current block, generating a second predictor for the current block, and combining the first compound predictor and the second predictor to obtain a prediction block for encoding or decoding the current block. An indicator may be sent from the encoder to the decoder to indicate when multi-level compound prediction is used to encode a block.
Abstract:
A method for encoding a video signal using a computing device, the video signal having a plurality of frames, each frame having a plurality of blocks, and each block having a plurality of pixels. The method includes generating a residual block from an original block of a current frame and a prediction block, degrading the residual block to decrease a bit-cost for encoding the residual block, and encoding the residual block into an encoded residual block.
Abstract:
Encoding and decoding using prediction dependent transform coding are provided. Encoding and decoding using prediction dependent transform coding may include identifying a current input block from a current input frame from an input video stream, generating a prediction block for the current input block, generating a residual block based on a difference between the current input block and the prediction block, generating, by a processor in response to instructions stored on a non-transitory computer readable medium, an encoded block by encoding the residual block based on the prediction block using the using prediction dependent transform coding, including the encoded block in an output bitstream, and outputting or storing the output bitstream.
Abstract:
An apparatus (e.g., a computer device) includes a video coding system. The video coding system includes an encoder. The encoder is configured to partition a block of video data into a first partition and a second partition, the first partition and the second partition being divided by a border that is at an angle to an orientation of a row or column of pixels, determine a first motion vector for the first partition and a second motion vector for the second partition, determine a first set of predictor values based on the first motion vector and a second set of predictor values based on the second motion vector, and combine the first set of predictor values and the second set of predictor values into an overall set of predictor values based on a weighting algorithm.
Abstract:
Restoring a degraded frame resulting from reconstruction of a source frame is described. A method includes generating, using first restoration parameters, a first guide tile for a degraded tile of the degraded frame, determining a projection parameter for a projection operation, and encoding, in an encoded bitstream, the first restoration parameters and the projection parameter. The projection operation relates differences between a source tile of the source frame and the degraded tile to differences between the first guide tile and the degraded tile.
Abstract:
Decoding a current block of an encoded video stream may include generating a base prediction block for the current block based on current prediction parameters associated with the current block, identifying adjacent prediction parameters used for decoding a previously decoded adjacent block that is adjacent to the current block, and determining an overlap region within the current block and adjacent to the adjacent block. The overlap region has a size being determined as a function of a difference between the first prediction parameters and the adjacent prediction parameters. For each pixel within the overlap region, an overlapped prediction of a pixel value may be generated as a function of the base prediction and a prediction based on the adjacent prediction parameters.