Abstract:
A multi-stage coding method includes receiving an input block of data for encoding and one or more previously coded samples associated with the input block. The input block is segmented into at least a first sub-region and a second sub-region. A prediction for the first sub-region is generated based on the one or more previously coded samples. Residual data for the first sub-region is obtained using the prediction for the first sub-region. A reconstruction of the first sub-region is generated using the residual data for the first sub-region and the prediction for the first sub-region. A prediction for the second sub-region is generated using the reconstruction of the first sub-region. Residual data for the second sub-region is obtained using the prediction for the second sub-region. The input block is encoded based in part on the residual data for the first region and the residual data for the second region.
Abstract:
A multi-stage coding method includes receiving an input block of data for encoding and one or more previously coded samples associated with the input block. The input block is segmented into at least a first sub-region and a second sub-region. A prediction for the first sub-region is generated based on the one or more previously coded samples. Residual data for the first sub-region is obtained using the prediction for the first sub-region. A reconstruction of the first sub-region is generated using the residual data for the first sub-region and the prediction for the first sub-region. A prediction for the second sub-region is generated using the reconstruction of the first sub-region. Residual data for the second sub-region is obtained using the prediction for the second sub-region. The input block is encoded based in part on the residual data for the first region and the residual data for the second region.
Abstract:
Image and video processing techniques are disclosed for processing components of a color space individually by determining limits for each component based on the relationship between each component in a color space. These limits may then be used to clip each component such that the component values are within the determined range for that component. In this manner, more efficient processing of images and/or video may be achieved.
Abstract:
Methods are described for encoding and decoding blocks of image data using intra block copying (IBC). A source block for intra block copying is selected from a source region of a current image that is closer to the current block than a threshold, wherein the source region does not include a portion of the current image that is further from the current block than the threshold.
Abstract:
A method and apparatus for variable accuracy inter-picture timing specification for digital video encoding is disclosed. Specifically, the present invention discloses a system that allows the relative timing of nearby video pictures to be encoded in a very efficient manner. In one embodiment, the display time difference between a current video picture and a nearby video picture is determined. The display time difference is then encoded into a digital representation of the video picture. In a preferred embodiment, the nearby video picture is the most recently transmitted stored picture. For coding efficiency, the display time difference may be encoded using a variable length coding system or arithmetic coding. In an alternate embodiment, the display time difference is encoded as a power of two to reduce the number of bits transmitted.
Abstract:
Scalable video coding and multiplexing compatible with non-scalable decoders is disclosed. In some embodiments, video data is received and encoded in a manner that renders at least a base layer to be compatible with a non-scalable video encoding standard, including by assigning for at least the base layer default values to one or more scalability parameters. In some embodiments, video data is received and encoded to produce an encoded video data that includes a base layer that conforms to a non-scalable video encoding standard and one or more subordinate non-scalable layers, which subordinate non-scalable layers do not by themselves conform to the non-scalable video encoding standard but which can he combined with the base layer to produce a result that does conform to the non-scalable video encoding standard, such that the result can be decoded by a non-scalable decoder. An identification data identifying those portions of the encoded video data that are associated with a subordinate non-scalable layer is included in the encoded video data.
Abstract:
A perimeter around a detected object in a frame of image data can be generated in a first coordinate system. The perimeter can be converted from the first coordinate system into a second coordinate system having the same aspect ratio as the first coordinate system. A first metadata entry can include dimensions of image data in the second coordinate system. A second metadata entry can provide a location and dimensions of the converted perimeter in the second coordinate space. Additional metadata can indicate matching objects between frames, position of an object relative to other objects in a frame, a probability that an object is correctly detected, and a total number of objects detected across multiple frames of image data.