Abstract:
A computing device generates a file that comprises a track box that contains metadata for a track in the file. Media data for the track comprises a sequence of samples, each of the samples being a video access unit of multi-layer video data. As part of generating the file, the computing device generates, in the file, a sub-sample information box that contains flags that specify a type of sub-sample information given in the sub-sample information box. When the flags have a particular value, a sub-sample corresponding to the sub-sample information box contains exactly one coded picture and zero or more non-Video Coding Layer (VCL) Network Abstraction Layer (NAL) units associated with the coded picture.
Abstract:
A video decoder receives a value for a first syntax element representing whether a dependency type syntax element for a current layer is signaled, wherein the dependency type syntax element identifies a type of dependency of a current layer relative to a reference layer; and in response to the value for the first syntax element indicating that the dependency type syntax element is not signaled determines that the type of dependency of the current layer relative to the reference layer is a predetermined type and decodes a block of the current layer using inter-layer prediction conforming to the predetermined type.
Abstract:
Techniques are described for sub-prediction unit (PU) based motion prediction for video coding in HEVC and 3D-HEVC. In one example, the techniques include an advanced temporal motion vector prediction (TMVP) mode to predict sub-PUs of a PU in single layer coding for which motion vector refinement may be allowed. The advanced TMVP mode includes determining motion vectors for the PU in at least two stages to derive motion information for the PU that includes different motion vectors and reference indices for each of the sub-PUs of the PU. In another example, the techniques include storing separate motion information derived for each sub-PU of a current PU predicted using a sub-PU backward view synthesis prediction (BVSP) mode even after motion compensation is performed. The additional motion information stored for the current PU may be used to predict subsequent PUs for which the current PU is a neighboring block.
Abstract:
In general, the disclosure relates to encoding and decoding a block of video data associated with three-dimensional (3D) video. A video coding device determines whether a depth condition associated with the block of video data should be set to valid within a coded bitstream. When the depth condition should be set to valid, the video coding device sets the depth condition to valid and encodes the block of video data using at least one camera parameter. The video coding device then determines whether the depth condition is valid. When the depth condition is valid, the video coding device decodes the block of video data using at least one camera parameter.
Abstract:
An example video coding device is configured to determine a depth value associated with a block of video data included in a dependent depth view, based on one or more neighboring pixels positioned adjacent to the block of video data in the dependent depth view, and generate a disparity vector associated with the block of video data, based at least in part on the determined depth value associated with the block of video data. The video coding device may further be configured to use the disparity vector to generate an inter-view disparity motion vector candidate (IDMVC), generate an inter-view predicted motion vector candidate (IPMVC) associated with the block of video data based on a corresponding block of video data in a base view, and determine whether to add any of the IDMVC and the IPMVC to a merge candidate list associated with the block of video data.
Abstract:
Techniques for advanced residual prediction (ARP) for coding video data may include inter-view ARP. Inter-view ARP may include identifying a disparity motion vector (DMV) for a current video block. The DMV is used for inter-view prediction of the current video block based on an inter-view reference video block. The techniques for inter-view ARP may also include identifying temporal reference video blocks in the current and reference views based on a temporal motion vector (TMV) of the inter-view reference video block, and determining a residual predictor block based on a difference between the temporal reference video blocks.
Abstract:
A device may determine, based on a value, whether all cross-layer random access skipped (CL-RAS) pictures of an intra random access point (IRAP) access unit are present in a video data bitstream. In addition, the device may reconstruct pictures of the video data based at least in part on syntax elements decoded from the video data bitstream.
Abstract:
Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving a non-entropy encoded layer dependency information at a position within a video parameter set (VPS) extension prior to syntax elements of the VPS extension that are entropy encoded. The systems, methods, and devices may encode or decode the non-entropy encoded layer dependency information before an entropy encoded syntax element. The systems, methods, and devices may encode or decode video data of one or more of the layers of video data based on the non-entropy encoded layer dependency information. The layer dependency information indicates whether one of the layers is a direct reference layer for another of the layers.
Abstract:
Systems, methods, and devices for coding multilayer video data are disclosed that may include encoding, decoding, transmitting, or receiving multilayer video data. The systems, methods, and devices may transmit or receive a video parameter set (VPS) including information for a series of layers, each layer including visual signal information. The systems, methods, and devices may code (encode or decode) video data based on the visual signal information signaled per layer in the VPS.
Abstract:
A video decoder performs a neighboring-block based disparity vector (NBDV) derivation process to determine a disparity vector or performs a NBDV refinement (NBDV-R) process to determine the disparity vector. The video decoder uses the disparity vector as a disparity vector for a current block without using a median filtering process on multiple disparity motion vectors, wherein the current block is coded in either a skip mode or a direct mode. Furthermore, the video coder determines pixel values for the current block.