Abstract:
Provided are systems and methods for processing 360-degree video data. In various implementations, a spherical representation of a 360-degree video frame can be segmented into a top region, a bottom region, and a middle region. The middle region can be mapped into one or more rectangular areas of an output video frame. The top region can be mapped into a first rectangular area of the output video frame using a mapping that converts a square to a circle, such that pixels in the circular top region are expanded to fill the first rectangular region. The bottom region can be mapped into a second rectangular area of the output video frame such that pixels in the circular bottom region are expanded to fill the second rectangular region.
Abstract:
Systems, methods, and computer-readable storage media are provided for decoded picture buffer (DPB) operations and rewriting access unit delimiters (AUDs) after bitstream extractions. An example method can include storing one or more pictures associated with an access unit (AU) in a decoded picture buffer (DPB), the AU including a first plurality of pictures, the first plurality of pictures corresponding to a plurality of video coding layers; after each picture of a second plurality of pictures associated with the AU is removed from a coded picture buffer (CPB), removing at least one picture of the one or more pictures from the DPB; and storing, in the DPB, each picture of the second plurality of pictures removed from the CPB.
Abstract:
Various embodiments may provide methods, systems, and devices for supporting application level discovery of Radio Access Network (RAN) statistics and/or events. Various embodiments may provide methods, systems, and devices for supporting application level signaling of Quality of Service (QoS) requirements.
Abstract:
The techniques of this disclosure allow for wavefront parallel processing of video data with limited synchronization points. In one example, a method of decoding video data comprises synchronizing decoding of a first plurality of video block rows at a beginning of each video block row in the first plurality of video block rows, decoding the first plurality of video block rows in parallel, wherein decoding does not include any synchronization between any subsequent video block in the first plurality of video block rows, and synchronizing decoding of a second plurality of video block rows at a beginning of each video block row in the second plurality of video block rows.
Abstract:
Systems and techniques are described herein for processing video data. For instance, a process can include obtaining video data including a picture. The process can also include determining a width of a film grain synthesis block of the picture based on at least one of a width and a height of the picture. The process can further include determining a height of the film grain synthesis block of the picture is one. The process may also include determining a block size of the film grain synthesis block based on the determined width and height. The process may further include selecting a grain block based on the determined block size.
Abstract:
Disclosed are systems, methods, apparatuses, and computer-readable media for processing video data. For instance, an apparatus for processing video data, may include: at least one memory; and at least one processor coupled to the at least one memory, the at least one processor being configured to: obtain a bitstream; retrieve a granularity type syntax element associated with the bitstream, the granularity type syntax element specifying a type of granularity of one or more pictures over which complexity metrics (CM) associated with the bitstream are applicable; retrieve a period type syntax element associated with the bitstream, the period type syntax element indicating an upcoming period of time or a set of pictures over which the CM are applicable; and decode a portion of the bitstream based on the granularity type syntax element and the period type syntax element.
Abstract:
Aspects of the present disclosure provide techniques for using a radio access network (RAN) level negative acknowledgement (NAK) feedback to indicate at least one missing frame from an encoding device. The RAN level NAK feedback replaces or preempts a decoding device sending an end-to-end feedback to the encoding device using real-time transport protocol (RTP) that has a long latency and may cause freezes at the decoding device. For example, an encoding device may send to a network entity a request for a configuration that configures the encoding device to transmit media frames to the decoding device. The network entity may provide a NAK feedback indicating at least one missing frame. Having received the configuration in response to the request, the encoding device transmits media frames to the decoding device via the network entity, and monitors for NAK feedback from the network entity in accordance with the configuration.
Abstract:
Techniques are described for video encoding and decoding using reference picture resampling with switchable filters. One example involves obtaining a current picture and a reference picture, identifying filter index information for a current block of the current picture, and determining that a first picture size value of the current picture and a second picture size value of the reference picture are different. Based on the determining that the first picture size value of the current picture and the second picture size value of the reference picture are different, performing a resampling process using a default filter index in place of a current filter index identified by the filter index information. Additional examples can use the current filter index identified by the filter index information in subsequent blocks. In various examples, the current filter index can be derived or signaled.
Abstract:
Techniques for encoding and decoding video data are described. A method of coding video may include determining a plurality of motion vector candidates for a block of video data for use in a motion vector prediction process, wherein each of the motion vector candidates points to a respective reference frame index, performing the motion vector prediction process using the motion vector candidates to determine a motion vector for the block of video data, and performing motion compensation for the block of video data using the motion vector and a common reference frame index, wherein the common reference frame index is used regardless of the respective reference frame index associated with the determined motion vector.
Abstract:
A video encoding device is configured to generate a first group of syntax elements. Each syntax element in the first group indicates whether a prediction mode of a respective prediction unit (PU) is based on an index into a list of most probable modes. A second group of syntax elements is generated that correspond to respective syntax elements in the first group. The syntax elements in the second group identify either an index into the list of most probable modes or an intra-prediction mode. The first group of syntax elements are context adaptive binary arithmetic coding (CABAC) encoded, and the second group of syntax elements are bypass encoded. A video decoder is configured to receive the entropy encoded first and second groups of syntax elements. The video decoder CABAC decodes the first group of flags and bypass decodes the second group of flags.