Abstract:
Aspects of this disclosure relate to a method of coding video data. In an example, the method includes obtaining, from an encoded bitstream and for any view component of a first view, reference view information indicating one or more reference views for predicting view components of the first view. The method also includes including, for decoding a first view component in an access unit and in the first view, one or more reference candidates in a reference picture list, where the one or more reference candidates comprise view components in the access unit and in the reference views indicated by the reference view information, where the number of reference candidates is equal to the number of reference views. The method also includes decoding the first view component based on the one or more reference candidates in the reference picture list.
Abstract:
A multiplexer may produce an MPEG-2 (Motion Picture Experts Group) System standard bitstream comprising views with nonconsecutive view order indexes. In one example, an apparatus includes a video encoder that encodes views of a scene, and a multiplexer that constructs a data structure for signaling that a corresponding MPEG-2 (Motion Picture Experts Group) System standard bitstream includes a first view of the views of the scene associated with a first view order index and a second view of the plurality of views of the scene associated with a second view order index. The first view order index and the second view order index are non-consecutive. The apparatus may further include an output interface that outputs the data structure.
Abstract:
In one example, a device for coding video data includes a video coder configured to code, for a bitstream, information representative of which of a plurality of video coding dimensions are enabled for the bitstream, and code values for each of the enabled video coding dimensions, without coding values for the video coding dimensions that are not enabled, in a network abstraction layer (NAL) unit header of a NAL unit comprising video data coded according to the values for each of the enabled video coding dimensions. In this manner, NAL unit headers may have variable lengths, while still providing information for scalable dimensions to which the NAL units correspond.
Abstract:
In one example, a video decoder is configured to store an adaptation parameter set (APS) data structure in a buffer of a coding device, wherein the APS data structure includes signaling data applicable to one or more slices of video data, decode at least one of the slices based on the signaling data of the APS data structure, after decoding the slice, determine that the APS data structure can be removed from the buffer, remove the APS data structure from the buffer based on the determination, and decode one or more additional slices of the video data after removing the APS data structure from the buffer.
Abstract:
In an example aspects of this disclosure generally relate to a method of coding video data that includes determining a first bit depth for outputting video data and a second bit depth for coding the video data, wherein the first bit depth is less than the second bit depth. The method also includes determining whether the video data will be used as reference data when coding other video data. The method also includes storing, based on the determination, the video data at the first bit depth when the video data is not used as reference data, and the video data at the second bit depth when the video data is used as reference data.
Abstract:
This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, a plurality of feature points may be determined for a 2D view. The plurality of feature points may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. For example, the plurality of feature points may be used to generate a global motion frame difference. The global motion frame difference may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation.
Abstract:
This disclosure describes techniques relevant to HTTP streaming of media data. According to these techniques, a server device may signal an open decoding refresh (ODR) random access point (RAP) for a movie segmentation of a movie representation. At least one frame of the media segmentation following the ODR RAP frame in decoding order may not be correctly decoded, and wherein each frame of the media segmentation following the ODR RAP frame in display order can be correctly decoded without relying on content of frames prior to the ODR RAP in display order. According to the techniques of this disclosure, a client device may communicate a request to a server device for the streaming of media data based on signaling of the ODR RAP. Also according to the techniques of this disclosure, a client device may commence decoding and/or playback of the movie representation based on signaling of the ODR RAP.
Abstract:
An asymmetric frame of a coded video bitstream may include a full resolution picture of a left view and a reduced resolution picture of a right view, where the left and right views form a stereo view pair for three-dimensional video playback. In one example, an apparatus includes a video encoder configured to receive a first picture of a first view of a scene having a first resolution, receive a second picture of a second view of the scene having a reduced resolution relative to the first resolution, form an asymmetric frame comprising the first picture and the second picture, and encode the asymmetric frame. In this manner, decoders of varying capabilities may receive the same bitstream, and the bitstream may consume less bandwidth than one or more bitstreams having full resolution pictures of a stereo view pair. The bitstream may have better quality than a bitstream having subsampled pictures.
Abstract:
In an example aspects of this disclosure generally relate to a method of coding video data that includes determining a first bit depth for outputting video data and a second bit depth for coding the video data, wherein the first bit depth is less than the second bit depth. The method also includes determining whether the video data will be used as reference data when coding other video data. The method also includes storing, based on the determination, the video data at the first bit depth when the video data is not used as reference data, and the video data at the second bit depth when the video data is used as reference data.
Abstract:
This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, a plurality of feature points may be determined for a 2D view. The plurality of feature points may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. For example, the plurality of feature points may be used to generate a global motion frame difference. The global motion frame difference may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation.