Abstract:
Techniques and systems are provided for encoding video data. For example, a method of encoding video data includes generating, by an encoder, an encoded video bitstream according to a first coding protocol. The encoded video bitstream includes one or more enhancement layers and a video parameter set defining parameters of the encoded video bitstream. The method further includes determining that a base layer is to be provided to a decoding device by an external source and is not to be provided as part of the encoded video bitstream. The method further includes assigning a minimum value to a syntax structure in the video parameter set when it is determined that the base layer is to be provided by the external source. The syntax structure defines profile, tier, and level parameters for the base layer.
Abstract:
A video decoder is configured to decode a bitstream that comprises an encoded representation of video data. As part of decoding the bitstream, the video decoder obtains, from the bitstream, one or more syntax elements indicating one or more partitioning schemes. For each respective partitioning scheme of the one or more partitioning schemes, the respective partitioning scheme specifies a respective set of disjoint partitions whose union forms an output layer set. Each respective partition of the respective set of disjoint partitions contains one or more of the layers. The video decoder is further configured to decode each of the partitions of a particular partitioning scheme using different processing cores in a plurality of hardware cores, the particular partitioning scheme being one of the one or more partitioning schemes.
Abstract:
Methods for coding recovery point supplemental enhancement information (SEI) messages and region refresh SEI messages in multi-layer coding are disclosed. In one aspect, the method may include determining at least one picture among a plurality of pictures included within an access unit with which to associate at least one recovery point SEI message, the access unit being included within the multi-layer bitstream. The method may further include associating the at least one recovery point SEI message with the determined at least one picture. The method may also include coding the video information based at least in part on the association of the at least one recovery point SEI message with the determined at least one picture.
Abstract:
In an example, a method of decoding video data includes decoding data that indicates a picture order count (POC) reset for a POC value of a first picture of a first received layer of multi-layer video data, where the first picture is included in an access unit. The method also includes, based on the data that indicates the POC reset for the POC value of the first picture and prior to decoding the first picture, outputting all pictures stored in a decoded picture buffer (DPB) that precede the first picture in coding order and that are not included in the access unit.
Abstract:
A computing device generates a file that comprises a track box that contains metadata for a track in the file. Media data for the track comprises a sequence of samples, each of the samples being a video access unit of multi-layer video data. As part of generating the file, the computing device generates, in the file, a sub-sample information box that contains flags that specify a type of sub-sample information given in the sub-sample information box. When the flags have a particular value, a sub-sample corresponding to the sub-sample information box contains exactly one coded picture and zero or more non-Video Coding Layer (VCL) Network Abstraction Layer (NAL) units associated with the coded picture.
Abstract:
An example method includes, receiving an encoded picture of 360-degree video data, the encoded picture of 360-degree video data being arranged in packed faces obtained from a projection of a sphere of the 360-degree video data; decoding the picture of encoded 360-degree video data to obtain a reconstructed picture of 360-degree video data, the reconstructed picture of 360-degree video data being arranged in the packed faces; padding the reconstructed picture of 360-degree video data to generate a padded reconstructed picture of 360-degree video data; in-loop filtering the padded reconstructed picture of 360-degree video data to generate a padded and filtered reconstructed picture of 360-degree video data; and storing the padded and filtered reconstructed picture of 360-degree video data in a reference picture memory for use in predicting subsequent pictures of 360-degree video data.
Abstract:
A device for determining information for video data includes one or more processors implemented in circuitry that are configured to determine one or more most-interested regions of a plurality of regions of an image of video data based on data representative of the one or more most-interested regions. The data representative of the one or more most-interested regions is external to video coding layer (VCL) data of a bitstream including the video data. The processors output a request for relatively high-quality video data for the one or more most-interested regions of the image and output, after outputting the request for the relatively high-quality video data, a request for relatively low-quality video data for one or more other regions of the image.
Abstract:
A device generates a Network Abstraction Layer (NAL) Unit Map Entry box in a file. The NAL Unit Map Entry box associates NAL units with a tile set that includes two or more tiles of a plurality of tiles into which a current picture is partitioned. A region of the current picture covered by the tile set may be non-rectangular. Additionally, the device generates a tile set group entry in the file. The tile set group entry includes a horizontal offset, a vertical offset, a width, and a height. The horizontal offset specifies a horizontal offset of a top-left pixel of a minimum rectangular region that covers tiles in the tile set. The vertical offset specifies a vertical offset of the top-left pixel of the minimum rectangular region. The width specifies a width of the minimum rectangular region. The height specifies a height of the minimum rectangular region.
Abstract:
A video coder may determine reference samples based on a location of a current block of a current picture of the 360-degree video data and a packing arrangement that defines an arrangement of a plurality of regions in the current picture. The current picture is in a projected domain and each respective region of the plurality of regions is a respective face defined by a projection of 360-degree video data. The regions are arranged in the current picture according to the packing arrangement. Based on the location of the current block being at a border of the first region that is adjacent to the second region and there being a discontinuity at the border due to the packing arrangement, the reference samples are samples of the current picture that spatially neighbor the current block in a spherical domain and not in the projected domain.
Abstract:
An example method includes, receiving an encoded picture of 360-degree video data, the encoded picture of 360-degree video data being arranged in packed faces obtained from a projection of a sphere of the 360-degree video data; decoding the picture of encoded 360-degree video data to obtain a reconstructed picture of 360-degree video data, the reconstructed picture of 360-degree video data being arranged in the packed faces; padding the reconstructed picture of 360-degree video data to generate a padded reconstructed picture of 360-degree video data; in-loop filtering the padded reconstructed picture of 360-degree video data to generate a padded and filtered reconstructed picture of 360-degree video data; and storing the padded and filtered reconstructed picture of 360-degree video data in a reference picture memory for use in predicting subsequent pictures of 360-degree video data.