Abstract:
Various implementations address depth coding and related disciplines. In one particular implementation, a segmentation is determined for a particular portion of a video image in a sequence of video images. The segmentation is determined based on reference depth indicators that are associated with at least a portion of one video image in the sequence of video images. Target depth indicators associated with the particular portion of the video image are processed. The processing is based on the determined segmentation in the particular portion of the video image. In another particular implementation, a segmentation is determined for at least a given portion of a video image based on depth indicators associated with the given portion. The segmentation is extended from the given portion into a target portion of the video image based on pixel values in the given portion and on pixel values in the target portion.
Abstract:
Methods and apparatus for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data, the original format selected from one of a sparse occlusion data format and a filled occlusion data format; arranging 2D data, which is associated with the occlusion data, at location 0 in a reference picture list; decoding the occlusion data to produce decoded occlusion data; when the indicator indicates the sparse occlusion data format for the occlusion data, filling a non-occlusion area of the occlusion data with data indicative of a defined characteristic to produce decoded occlusion data; when the indicator indicates the filled occlusion data format for the occlusion data, replacing the occlusion data in a non-occlusion area of the occlusion data with sample values from a corresponding area of associated 2D data to produce decoded occlusion data; and outputting the decoded occlusion.
Abstract:
A method for processing video information by a video encoder includes classifying video information based on at least one local directional feature of the video information to design multiple filters. The encoder encodes filter coefficients of the multiple filters. The multiple filters are designed with symmetric constraints on the filter coefficients configured according to the classification of the video information based on at least one local directional feature of the video information. The method also includes applying, by the encoder, the multiple directional filters to the video information. A decoder is configured to decode filter coefficients of multiple filters and apply the multiple filters to the video information. The decoder constructs, from decoded filter coefficients, multiple filters with symmetric constraints on the filter coefficients configured according to the classification of the video information based on at least one local directional feature of the video information.
Abstract:
Various implementations are directed to providing inter-layer dependency information. In one implementation, syntax elements are generated that indicate an inter-layer dependency structure among three-dimensional video (3DV) layers. Based on the inter-layer dependency structure, an inter-layer reference is identified for a picture from a layer of the 3DV layers. The picture is encoded based, at least in part, on the inter-layer reference. Corresponding decoding implementations are also provided. Additionally, in another implementation, a transmission priority and an indication of network congestion are used to determine whether to transmit data for a particular 3DV layer. The transmission priority is based on an inter-layer dependency structure among multiple 3DV layers. Another implementation is directed to a network abstraction layer unit that can explicitly identify and convey inter-layer references and corresponding dependencies.
Abstract:
A method of video coding using palette coding mode for color video in a non-444 color format is disclosed. The non-444 video data can be converted to the 444 format so that the palette designed for the 444 format can be used for the non-444 video data. For the 4:2:0 color format, each chroma sample collocated with corresponding quad luma samples is padded with three chroma samples to form converted quad chroma samples. For the 4:2:0 color format, if a current pixel does not belong to the color palette, one or more escape values are signaled for the current pixel and the escape value associated with the chroma sample of the current pixel for each color component is not signaled for three of the quad chroma sample locations collocated with corresponding quad luma samples.
Abstract:
Methods and apparatuses for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data, the original format selected from one of a sparse occlusion data format and a filled occlusion data format; decoding the received occlusion data to produce decoded occlusion data; and when the indicator indicates the original format as a filled occlusion data format, converting the decoded occlusion data from a sparse occlusion data format to the filled occlusion data format, the converting further including; replacing non-occlusion area data, which is represented with a defined characteristic, by respective collocated samples from 2D data in the video data frame associated with the occlusion data; outputting the decoded occlusion data and, when present, converted decoded occlusion data.
Abstract:
Methods and apparatus for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data, the original format selected from one of a sparse occlusion data format and a filled occlusion data format; arranging 2D data, which is associated with the occlusion data, at location 0 in a reference picture list; decoding the occlusion data to produce decoded occlusion data; when the indicator indicates the sparse occlusion data format for the occlusion data, filling a non-occlusion area of the occlusion data with data indicative of a defined characteristic to produce decoded occlusion data; when the indicator indicates the filled occlusion data format for the occlusion data, replacing the occlusion data in a non-occlusion area of the occlusion data with sample values from a corresponding area of associated 2D data to produce decoded occlusion data; and outputting the decoded occlusion.
Abstract:
A method and apparatus for palette prediction and sharing according to the present invention are disclosed. A method incorporating an embodiment of the present invention determines one or more palette sharing flags for the current block. A set of current palettes corresponding to the set of color components is generated according to the palette sharing flags. If a first palette sharing flag is asserted, one or more current palettes indicated by the first palette sharing flag are copied entirely from one or more reference palettes of a set of reference palettes. If the first palette sharing flag is not asserted, one or more current palettes indicated by the first palette sharing flag are derived from a bitstream associated with the video data. Encoding or decoding is then applied to the current block according to the set of current palettes.
Abstract:
Methods and apparatuses for coding occlusion layers, such as occlusion video data and occlusion depth data in 3D video, are disclosed. A decoding method comprising the steps of: extracting an indicator representative of an original format for received occlusion data , the original format selected from one of a sparse occlusion data format and a filled occlusion data format; decoding the received occlusion data to produce decoded occlusion data ; and when the indicator indicates the original format as a filled occlusion data format, converting the decoded occlusion data from a sparse occlusion data format to the filled occlusion data format , the converting further including; replacing non-occlusion area data, which is represented with a defined characteristic, by respective collocated samples from 2D data in the video data frame associated with the occlusion data ; outputting the decoded occlusion data and, when present, converted decoded occlusion data .
Abstract:
Several implementations relate to 3D video (3DV) coding formats. One implementation encodes multiple pictures that describe different three-dimensional (3D) information for a given view at a given time. Syntax elements are generated that indicate, for the encoded multiple pictures, how the encoded picture fits into a structure that supports 3D processing. The structure defines content types for the multiple pictures. A bitstream is generated that includes the encoded multiple pictures and the syntax elements. The inclusion of the syntax elements provides, at a coded-bitstream level, indications of relationships between the encoded multiple pictures in the structure. The syntax elements also enable efficient inter-layer coding of the 3DV content, thereby reducing the bandwidth used to transmit the 3DV content. Corresponding decoding implementations are also provided. Extraction methods are also provided from extracting pictures of interest from such a 3DV multiple pictures and the syntax elements, the video stream characterized by such a 3D structure.