Abstract:
A system and method for regenerating high dynamic range (HDR) video data from encoded video data, extracts, from the encoded video data, a self-referential metadata structure specifying a video data reshaping transfer function. The video data reshaping transfer function is regenerated using data from the metadata structure and the extracted reshaping transfer function is used to generate the HDR video data by applying decoded video data values to the reshaping transfer function.
Abstract:
A method of encoding a digital video data applies adaptive pre-processing to data representing high dynamic range (HDR) and/or wide color gamut (WCG) image data prior to encoding and complementary post-processing to the data after decoding in order to allow at least partial reproduction of the HDR and/or WCG data. The example methods apply one or more color space conversions, and a perceptual transfer functions to the data prior to quantization. The example methods apply inverse perceptual transfer functions and inverse color space conversions after decoding to recover the HDR and/or WCG data. The transfer functions are adaptive so that different transfer functions may be applied to video data sets including different groups of frames, frames or processing windows in a single frame. Information on the data set and information on the applied transfer function is passed as metadata from the encoder to the decoder.
Abstract:
A method of encoding a digital video, comprising receiving a high dynamic range (HDR) master of a video, a reference standard dynamic range (SDR) master of the video, and target SDR display properties at an encoder, finding a color volume transform that converts HDR values from the HDR master into SDR values that, when converted for display on the target SDR display, are substantially similar to SDR values in the reference SDR master, converting HDR values from the HDR master into SDR values using the color volume transform, generating metadata items that identifies the color volume transform to decoders, and encoding the SDR values into a bitstream.
Abstract:
A video data encoder generates a first metadata structure to describe one or more transfer functions to be applied by a decoder to reshape decoded video data into video data. The encoder segments a transfer function relative to P pivot points. It then allocates P cells in the metadata structure for storing the P pivot points. Each transfer function segment is fitted to a respective equation having N coefficients based on the order of the equation. The encoder allocates N+1 cells for each pivot point except for the P th pivot point and stores the number N in the first cell and the N coefficients in the remaining cells. The encoder generates a second metadata structure, associated with a video data set, that includes data identifying the transfer function to be applied to the video data set. The encoder encodes the video data including the first and second metadata structures.
Abstract:
A method is provided for encoding a digital video to provide for improved color mapping. The digital video has values in a first color space, and the method includes performing a color mapping operation on values in each sub-picture to convert the values in the first color space to values in a second, narrower, color space, wherein the color mapping operation is adapted based on the content of each sub-picture, encoding the values in the second color space into a base layer, performing a reverse color mapping operation on decoded values from the base layer in the second color space in each sub-picture to generate a reconstructed reference frame having values in the first color space, encoding values in the first color space into an enhancement layer based at least in part on the reconstructed reference frame, combining the base layer and the enhancement layer into a bitstream, sending the bitstream to a decoder, and sending one or more parameters to the decoder that describe the adaption of the reverse color mapping operation for at least some sub-pictures.
Abstract:
A method of decoding JVET video, comprising receiving a bitstream indicating how a coding tree unit was partitioned into coding units according to a partitioning structure that allows nodes to be split according to a partitioning technique. An intra direction mode for a coding unit may be selected, as well as one or more of the plurality of reference lines to generate at least one predictor for the intra direction mode. A predictor may be generated from reference samples within each selected reference line by combining predicted pixel values based on a projected position on a main reference line in combination with predicted pixel values based on a projected position on a side reference line. The predicted pixel values are weighted according to a weight parameter, wherein the weight parameter is determined based on a shift conversion factor.
Abstract:
A method for transforming high dynamic range (HDR) video data into standard dynamic range (SDR) video data and encoding the SDR video data so that the HDR video data may be recovered at the decoder includes generating a tone map describing a transformation applied to the HDR video data to generate the SDR video data. The generated tone map describes the transformation as the multiplication of each HDR pixel in the HDR video data by a scalar to generate the SDR video data. The tone map is then modeled as a reshaping transfer function and the HDR video data is processed by the reshaping transfer function to generate the SDR video data. The reshaping transfer function is then inverted and described in a self-referential metadata structure. The SDR video data is then encoded including the metadata structure defining the inverse reshaping transfer function.
Abstract:
A method is provided for encoding a digital video to improve perceptual quality. The method includes receiving a digital video at a video encoder, providing a perceptual quantizer function defined by Formula (I); wherein L is a luminance value, c 1 , c 2 , c 3 , and m 1 are parameters with fixed values, and m 2 is a parameter with a variable value, adapting the perceptual quantizer function by adjusting the value of the m 2 parameter based on different luminance value ranges found within a coding level of the digital video, encoding the digital video into a bitstream using, in part, the perceptual quantizer function, transmitting the bitstream to a decoder, and transmitting the value of the m 2 parameter to the decoder for each luminance value range in the coding level.
Abstract:
A method of decoding JVET video, comprising defining a coding unit (CU) template within a decoded area of a video frame, the CU template being positioned above and/or to the left of a current decoding position for which data was intra predicted, defining a search window within the decoded area, the search window being adjacent to the CU template, generating a plurality of candidate prediction templates based on pixel values in the search window, each of the plurality of candidate prediction templates being generated using different intra prediction modes, calculating a matching cost between the CU template and each of the plurality of candidate prediction templates, selecting an intra prediction mode that generated the candidate prediction template that had the lowest matching cost relative to the CU template, and generating a prediction CU for the current decoding position based on the intra prediction mode.
Abstract:
A method of decoding a bitstream (108) comprises decoding the bitstream into color values and metadata items indicating information about adaptive post-processing operations performed by a decoder (302), performing high dynamic range (HDR) adaptation operations on the color values based on the metadata items (308), and performing fixed post-processing operations to reconstruct an HDR video from the color values (312), wherein the HDR adaptation operations convert color values into a format expected by the fixed post-processing operations.