Abstract:
In one example, a device for coding video data includes a video coder configured to determine, for a depth block of a depth component of video data, a co-located texture block of a corresponding texture component, and when at least a portion of the texture block corresponds to a prediction unit of the texture component that is not intra-prediction coded: disable an inter-component Wedgelet depth modeling mode for the depth block, select an intra-prediction coding mode for the depth block other than the disabled inter-component Wedgelet depth modeling mode, and code the depth block using the selected intra-prediction coding mode.
Abstract:
In one example, an apparatus includes a processor configured to provide information to a client device for accessing data for a first representation of a video sequence and a second representation of the video sequence, and to provide information to the client device indicating that the client device can, during transmission of the data for the first representation, switch to the second representation without experiencing a presentation gap when displaying the video sequence and without simultaneously executing more than one decoder to decode the data for the first representation and the second representation during the switch, in response to a first request from the client device to retrieve data from the first representation. In this manner, the client device may use the information to perform seamless switching between the representations.
Abstract:
In one example, a video coder is configured to code information indicative of whether view synthesis prediction is enabled for video data. When the information indicates that view synthesis prediction is enabled for the video data, the video coder may generate a view synthesis picture using the video data and code at least a portion of a current picture relative to the view synthesis picture. The at least portion of the current picture may comprise, for example, a block (e.g., a PU, a CU, a macroblock, or a partition of a macroblock), a slice, a tile, a wavefront, or the entirety of the current picture. On the other hand, when the information indicates that view synthesis prediction is not enabled for the video data, the video coder may code the current picture using at least one of intra-prediction, temporal inter-prediction, and inter-view prediction without reference to any view synthesis pictures.
Abstract:
A video coding device may be configured to code a bitstream including multiple views plus depth information. Two of the views may have reduced resolutions, while a third view may have a full resolution. The third view may be predicted relative to upsampled versions of the two reduced-resolution views. Each view may include texture data and depth data, such that a view component may include a texture component and a depth component. Moreover, the texture and depth components may be arranged within an access unit according to a particular order, which may simplify component extraction from the access unit.
Abstract:
A video coder performs a padding operation that processes a set of border pixels according to an order. The order starts at a bottom-left border pixel and proceeds through the border pixels sequentially to a top-right border pixel. When the padding operation processes an unavailable border pixel, the padding operation predicts a value of the unavailable border pixel based on a value of a border pixel previously processed by the padding operation. The video coder may generate an intra-predicted video block based on the border pixels.
Abstract:
Implementations include methods and systems for a converting reference images or video to 3D images or video. A two-step conversion is described which accomplishes warping and hole filling on a pixel-by-pixel basis. In one implementation, of a plurality of pixel values of a reference image at a plurality of first collinear pixels locations are successively mapped to a respective plurality of second pixel locations of a destination image. Between two of the mappings, a location of a hole between two of the second pixel locations may be identified and filled.
Abstract:
Aspects of this disclosure relate to a method of coding video data. In an example, the method includes determining a first residual quadtree (RQT) depth at which to apply a first transform to luma information associated with a block of video data, wherein the RQT represents a manner in which transforms are applied to luma information and chroma information. The method also includes determining a second RQT depth at which to apply a second transform to the chroma information associated with the block of video data, wherein the second RQT depth is different than the first RQT depth. The method also includes coding the luma information at the first RQT depth and the chroma information at the second RQT depth.
Abstract:
This disclosure describes techniques for coding 3D video block units. In one example, a video encoder is configured to receive one or more texture components from at least a portion of an image representing a view of three dimensional video data, receive a depth map component for at least the portion of the image, code a block unit indicative of pixels of the one or more texture components for a portion of the image and the depth map component. The coding comprises receiving texture data for a temporal instance of a view of video data, receiving depth data corresponding to the texture data for the temporal instance of the view of video data, and encapsulating the texture data and the depth data in a view component for the temporal instance of the view, such that the texture data and the depth data are encapsulated within a common bitstream.
Abstract:
This disclosure relates to techniques for constructing a combined reference picture list, List C, based on List 0 and List 1, used for uni-directional prediction of video blocks in any direction. The techniques include coding one or more syntax elements defined to indicate construction information for List C, and performing reference picture list construction for List C from List 0 and List 1 based on the syntax elements. The one or more syntax elements may indicate that List C is used for uni-directional prediction, and may also indicate a number of reference pictures identified in List C and a reference index of a reference picture for each entry in List C. Each coded video block of a B slice may have an associated syntax element, i.e., inter_pred_idc, to indicate whether the video block is bi-predicted from List 0 and List 1 (Bi) or uni-directional predicted from List C (Pred_LC).
Abstract:
This disclosure describes techniques for estimating a depth of image objects for a two-dimensional (2D) view of a video presentation. For example, an initial indication of depth (e.g., an optical flow) may be determined for a 2D view. The initial indication of depth may be used to estimate global motion, e.g., motion of an observer (e.g., camera), of the 2D view. The initial indication of depth may be modified based on the estimation of global motion to create a global motion-adjusted indication of depth. The global motion-adjusted depth indication may be used to create a depth map for the 2D view, which may be used to generate an alternative view of the video presentation that may be used to display a three-dimensional (3D) video presentation.