Abstract:
Video coding infrastructure using adaptive prediction complexity reduction. One or more subsets associated with one or more frames or pictures of the video signal may be adaptively selected and used for motion vector calculation (e.g., such as in accordance with inter-prediction). For example, a picture or frame of the video signal may be partitioned into a number of respective regions. Any one or more, but typically fewer than all, of the respective regions may be appropriately selected, and stored, based on any one or more considerations for use in motion vector calculation (e.g., inter-prediction). A sub-sampled or down-sampled picture or frame [or alternatively, a sub-sampled or down-sampled version of one or more respective regions of a picture or frame] (e.g., the sub-sampling or down-sampling ratio which may be adaptively determined based on any one or more considerations) may be stored for use in motion vector calculation (e.g., inter-prediction).
Abstract:
Multi-mode error concealment, recovery and resilience coding. Adaptation of a number of coding units (CUs) employed in accordance with video coding may be made as a function of error. As a number of errors increases, the respective number of CUs may correspondingly increase (e.g., which may be made in accompaniment with a reduction of CU size). As a number of errors decreases, the respective number of CUs may correspondingly decrease (e.g., which may be made in accompaniment with an increase of CU size). Such errors may be associated with a type of source providing a video signal, a type of error resilience coding employed, communication link and/or channel conditions, a remote error characteristic (e.g., such as associated with a source device and/or destination device), a local error characteristic (e.g., such as associated with operations and/or processing within a given device), and/or any other type of consideration.
Abstract:
Methods and systems for processing B pictures with missing or invalid forward reference pictures are disclosed. Aspects of one method may include decoding of a contiguous sequence of B pictures that immediately follow a first I picture after a sequence header. The decoding of the B pictures may handle video edit codes in the AVS1-P2 format and/or random access points. Some embodiments of the invention may discard B pictures that refer to invalid or missing forward reference pictures, while other embodiments of the invention may interpolate B pictures. For an invalid forward reference picture, the interpolation algorithm for generating pictures may use the invalid forward reference picture and the backward reference picture. For a missing decoded forward reference picture, as in a random access, the interpolation algorithm for generating pictures may use the backward reference picture.
Abstract:
Aspects of a method and system for processing audio and video transport may comprise synchronizing streams with associated program content based on 24-bit presentation time stamps (PTSs) and 32-bit system clock samples (SCSs) field without using decoding time stamps (DTSs). A PTS field may be provided as part of a packetized elementary stream (PES) packet syntax. An SCS field may be provided as part of a transport stream (TS) syntax. A system clock (STC) may be determined from received SCS values. Video and/or audio data may be decoded as soon as they arrive and when there is at least one display and/or presentation frame buffer available. The video and/or audio data may be presented for display and/or play based on a comparison of the PTS and STC values. When a frame does not have an associated PTS, a PTS extrapolation operation may generate an extrapolated PTS value from previously received PTSs.
Abstract:
Various methods and systems are provided for latency aware priority based decoding. In one embodiment, among others, a method includes providing coded frames of a first input stream to a multiple channel (multi-channel) decoder for decoding. A second input stream is obtained and a latency priority associated with the second input stream is determined. Coded frames from the first and second input streams are provided to the multi-channel decoder for decoding in an order based upon the latency priorities associated with the first and second input streams. In another embodiment, a multi-channel decoding system includes a multi-channel decoder configured to decode coded frames and a decoder input scheduler configured to provide coded frames from a plurality of input streams to the multi-channel decoder for decoding. The coded frames are provided to the multi-channel decoder in an order based at least in part upon latency priorities associated with the input streams.
Abstract:
Methods and systems for processing B pictures with missing or invalid forward reference pictures are disclosed. Aspects of one method may include decoding of a contiguous sequence of B pictures that immediately follow a first I picture after a sequence header. The decoding of the B pictures may handle video edit codes in the AVS1-P2 format and/or random access points. Some embodiments of the invention may discard B pictures that refer to invalid or missing forward reference pictures, while other embodiments of the invention may interpolate B pictures. For an invalid forward reference picture, the interpolation algorithm for generating pictures may use the invalid forward reference picture and the backward reference picture. For a missing decoded forward reference picture, as in a random access, the interpolation algorithm for generating pictures may use the backward reference picture.
Abstract:
Advance encode processing based on raw video data. Adaptive processing of a video signal may be made by making one or more decisions and/or directing one or more encoding steps based upon characteristic(s) associated with the video signal in raw form (e.g., before the video signal undergoing any processing in accordance with generating an output video bitstream). Characteristic(s) associated with the video signal (raw form) may be used to drive intra-prediction angular mode selection, including making a preliminary/coarse angular mode selection (e.g., which may be refined and/or modified based upon partial and/or full processing of the video signal). Characteristic(s) associated with the video signal (raw form) may be used to drive intra-prediction/inter-prediction operational mode selection (or intra-prediction/inter-prediction/neither operational mode selection). An integrated and/or embedded video decoder within a video encoding device may be provisioned to assist in adaptation of one or more of the video encoding operations.
Abstract:
Video coding tree sub-block splitting. A parser and treeblock splitter is implemented for adaptive partitioning of treeblocks (TBs) into various respective sub-treeblocks (STBs). Such adaptation may be tailored for effectuating parallel processing in accordance with video decoding. Different respective decoding engines may each individually and respectively also perform further partitioning of the STBs into sub-STBs as well. Such adaptation of respective STB size and/or sub-STB size may be made based on local and/or remote consideration(s). For example, adaptation may be made based upon communication link and/or channel conditions, a remote characteristic (e.g., associated with a source device and/or destination device), a local characteristic (e.g., associated with operations and/or processing performed locally within a given device), and/or any other type of consideration. Different respective portions of a video signal may be processed using different respective sized STBs/sub-STBs (and different respective engines may also employ different respective sized STBs/sub-STBs).
Abstract:
Within a device, one or more subsets associated with one or more frames or pictures of the video signal may be adaptively selected and used for motion vector calculation (e.g., such as in accordance with inter-prediction). For example, a picture or frame of the video signal may be partitioned into a number of respective regions. Any one or more, but typically fewer than all, of the respective regions of that picture or frame may be appropriately selected, and stored, based on any one or more considerations for use in motion vector calculation (e.g., inter-prediction). A sub-sampled or down-sampled picture or frame [or alternatively, a sub-sampled or down-sampled version of one or more respective regions of a picture or frame] (e.g., the sub-sampling or down-sampling ratio which may be adaptively determined based on any one or more considerations) may be stored for use in motion vector calculation (e.g., inter-prediction).
Abstract:
Video coding sub-block sizing based on infrastructure capabilities and current conditions. Sub-block size, such as employed in accordance with the video processing, maybe adaptively modified based on any of a number of considerations. For example, such adaptation of sub-block size may be made with respect to one or more characteristics associated with streaming media source flow(s) and/or streaming media delivery flow(s) being received by and/or output from a given device including a video processor. For example, such a video processor may be a video decoder implemented within a middling or destination device. Such a video processor may be a video encoder implemented within the middling or source device. Adaptation of sub-block size employed in accordance with video coding may also be effectuated in accordance with feedback or control signaling provided between respective devices. (e.g., from destination or source device to middling device, or from destination device to source device, etc.).