Abstract:
There is provided a video signal processing apparatus, including: a first interface configured to obtain a first video signal imaged within a first dynamic range; a second interface configured to obtain information for specifying a partial area of the obtained first video signal; and an arithmetic processing circuit configured to calculate a representative value of brightness of the first video signal of the partial area, to compress a dynamic range of the first video signal depending on the calculated representative value, and to generate a second video signal, a dynamic range of the second video signal being restricted to a second dynamic range narrower than the first dynamic range.
Abstract:
A method for encoding at least one depth image frame within a sequence of depth image frames, the method comprising: segmenting each at least one depth image frame into an array of depth image blocks;determining for each depth image block a minimum depth image value and a maximum depth image value;generating a maximum depth image from the determined maximum depth image values;generating a minimum depth image from the determined minimum depth image values; and encoding values of the at least one depth image frame based on values of the maximum depth image and values of the minimum depth image at corresponding positions.
Abstract:
Given HDR and SDR video inputs representing the same content, segment-based methods are described to generate a backward-compatible reshaped SDR video which preserves the artistic intent or "look" of the inputs and satisfies other coding requirements. For each frame in a segment, reshaping functions are generated based on a support frames set determined based on a sliding window of frames that is adjusted based on scene cuts in the segment and which may include frames from both the current segment and neighboring segments. For luma reshaping, a mapping that preserves the cumulative density function of the luminance histogram values in the EDR and SDR inputs is combined with a minimum codeword allocation derived based on the EDR signal and the support frame set. For chroma reshaping, methods for segment-based forward and backward reshaping using multivariate, multi-regression models are also presented.
Abstract:
Methods and systems for adaptive chroma reshaping are discussed. Given an input image, a luma-reshaped image is first generated based on its luma component. For each chroma component of the input image, the range of the pixel values in the luma reshaped image is divided into bins, and for each bin a maximal scale factor is generated based on the chroma pixel values in the input image corresponding to the pixels of the luma reshaped image in the bin. A forward reshaping function is generated based on a reference reshaped function and the maximal scale factors, and reshaped chroma pixel values for the chroma component are generated based on the forward reshaping function and the corresponding pixel values in the luma reshaped image. Implementations options using look-up tables for mobile platforms with limited computational resources are also described.
Abstract:
Systems and methods are described for enabling a client device to request video streams with different bit depth remappings for different viewing conditions. In an embodiment, information indicating the availability of additional remapped profiles is sent in a manifest file. Alternative bit-depth remappings may be optimized for different regions of interest in the image or video content, or for different viewing conditions, such as different display technologies and different ambient illumination. Some embodiments based on the DASH protocol perform multiple depth mappings at the encoder and also perform ABR-encoding for distribution. The manifest file contains information indicating additional remapping profiles. The remapping profiles are associated with different transformation functions used to convert from a higher bit-depth to a lower bit-depth.
Abstract:
A system and method for regenerating high dynamic range (HDR) video data from encoded video data, extracts, from the encoded video data, a self-referential metadata structure specifying a video data reshaping transfer function. The video data reshaping transfer function is regenerated using data from the metadata structure and the extracted reshaping transfer function is used to generate the HDR video data by applying decoded video data values to the reshaping transfer function.
Abstract:
The present principles are directed to a parameterized OETF/EOTF for processing images and video. The present principles provide a method for encoding a picture, comprising: applying a parameterized transfer function to a luminance (L) signal of the picture to determine a resulting V(L) transformed signal; encoding the resulting V(L); wherein the parameterized transfer function is adjusted based on a plurality of parameters to model one of a plurality of transfer functions. The present principles also provide for a method for decoding a digital picture, the method comprising: receiving the digital picture; applying a parameterized transfer function to the digital picture to determine a luminance (L) signal of the digital picture, the parameterized transfer function being based on a plurality of parameters; wherein the parameterized transfer function is adjusted based on a plurality of parameters to model one of a plurality of transfer functions.
Abstract:
The present disclosure generally relates to a method and device of encoding a color picture having color components (Ec), characterized in that it comprises: - obtaining (130) a luminance component (L) comprising: - obtaining (120) a modulation value (Ba) from the luminance (Y) of the color picture; - obtaining a scaled luminance by dividing the luminance (Y) of the color picture by said modulation value (Ba); - obtaining the luminance component (L) by applying a non-linear function on said scaled luminance in order that the dynamic of said luminance component (L) is reduced compared to the dynamic of said scaled luminance; - obtaining two chrominance components (C1, C2) comprising: - obtaining a factor (r(L(i)) that depends on the value of the pixel (i) of said luminance component (L(i)) and the luminance value (Y(i)) of the co-located pixel (i) in the color picture; - obtaining (150) at least one intermediate color component (E'c) by multiplying each color component (Ec) by said factor (r(L(i)); and - obtaining (170) said two chrominance components (C1, C2) from said at least one intermediate color components (E'c); and - encoding (180) said luminance (L) and two chrominance components (C1, C2).
Abstract:
Provided is an image processing apparatus and an image processing method. The image processing apparatus includes circuitry configured to receive image data, and receive metadata set based on the image data. The metadata including information serving as a display mapping reference to maintain a characteristic of the image data to a portion of the image data.
Abstract:
A video coding/decoding system codes data efficiently even when input video data exhibits changes in dynamic range. The system may map pixel values of the first frame from a dynamic range specific to the input image data to a second dynamic range that applies universally to a plurality of frames that have different dynamic ranges defined for them. The system may code the mapped pixel values to reduce bandwidth of the mapped frame data, and thereafter transmit the coded image data to a channel.