Abstract:
In one implementation, a method of generating a bit stream encoding a video stream is performed by a device including one or more processors and non-transitory memory. The method includes decomposing a video stream into a plurality of frequency band video streams. The method includes determining a target bitrate and determining, for each frequency band video stream, a respective frequency band bit rate based on the target bit rate. The method includes encoding each of the plurality of frequency band video streams at its respective frequency band bit rate and transmitting, over a channel, each encoded frequency band video stream.
Abstract:
Techniques are disclosed for coding video data in which frames from a video source are partitioned into a plurality of tiles of common size, and the tiles are coded as a virtual video sequence according to motion-compensated prediction, each tile treated as having respective temporal location of the virtual video sequence. The coding scheme permits relative allocation of coding resources to tiles that are likely to have greater significance in a video coding session, which may lead to certain tiles that have low complexity or low motion content to be skipped during coding of the tiles for select source frames. Moreover, coding of the tiles may be ordered to achieve low coding latencies during a coding session.
Abstract:
Techniques for encoding video with temporal layering are described, comprising predicting a sequence of pictures with a motion prediction reference pattern having a number of virtual temporal layers, and encoding the sequence of pictures into an encoded bitstream with a temporal layering syntax, wherein a number of signaled temporal layers is less than the number of virtual temporal layers. The number of signaled temporal layers may be determined from a target highest frame rate, a target base layer frame rate, and the number of virtual temporal layers.
Abstract:
Techniques are disclosed for developing quantization matrices for use in video coding. According to these techniques a first quantization matrix may be derived from a second quantization matrix by scaling quantization values of the second quantization matrix by scaling parameters. The scaling parameters may increase according to distance between each matrix position and a matrix origin, they may be derived from characteristics of a video sequence to be coded, or both. The first quantization matrix may be communicated to a decoder. Thereafter, a video sequence may be coded predictively. As part of the coding, pixel data of the video sequence may be transformed to a plurality of frequency domain coefficients, and the frequency domain coefficients may be quantized according to the first quantization matrix.
Abstract:
This disclosure relates to a wide gamut encoder capable of receiving a wide gamut color image in accordance with a wide gamut standard. The encoder can encode one or more wide gamut color image pixel values into portions of narrow gamut encoding elements for transmission to a video encoder. The encoder can implement an advanced extended YCC format that is backward compatible with a P3 color gamut.
Abstract:
Methods are described for encoding and decoding blocks of image data using intra block copying (IBC). A source block for intra block copying is selected from a source region of a current image that is closer to the current block than a threshold, wherein the source region does not include a portion of the current image that is further from the current block than the threshold.
Abstract:
This disclosure describes techniques to concurrently capture several redundant (e.g., overlapping) video image streams on electronic devices having multiple image capture devices (e.g., on mobile devices having two or more digital video cameras). The techniques described herein advantageously capture at least one “compressed” video image stream, such that the overall multi-video image stream capture process uses significantly less power than capturing each stream independently in an “uncompressed” fashion. These techniques provide an electronic device with the ability to provide a user with more video streams that are of a higher frame rate and/or higher quality for a longer capture time-without exceeding the device's thermal limits. In particular, a “compressed” amount of image information may be processed and saved from at least one camera at capture time, and then deferred processing operations may be used to reconstruct any missing information and produce the higher frame rate/higher quality video image streams.
Abstract:
A device comprises memory, a display characterized by a display characteristic, and processors coupled to the memory. The processors execute instructions causing the processors to receive data indicative of the display characteristic, data indicative of ambient lighting, and data indicative of content characteristics for a content item; determine a tone mapping curve for the content item based on the data indicative of content characteristics; determine a first, so-called “anchor” point along the tone mapping curve; modify a first portion of the tone mapping curve below the anchor point based on the data indicative of ambient lighting; modify a second portion of the tone mapping curve above the anchor point based on the data indicative of the display characteristic; perform tone mapping for the content item based on the modified toned mapping curve to obtain a tone mapped content item; and cause the display to display the tone mapped content item.
Abstract:
Methods are described for encoding and decoding blocks of image data using intra block copying (IBC). A source block for intra block copying is selected from a source region of a current image that is closer to the current block than a threshold, wherein the source region does not include a portion of the current image that is further from the current block than the threshold.
Abstract:
Methods and systems provide an adaptive quantization parameter (QP) modulation scheme for video coding and compression that is sensitive to user visual perception. In an embodiment, the method includes detecting an eye sensitive region, where a region is considered sensitive based on a noticeability of a visual effect. The method includes estimating encoding parameters for image content in the detected eye sensitive region. The method further includes encoding the detected eye sensitive region with the estimated encoding parameters. The estimating the encoding parameters may be based on, among other things, a variance, a motion vector, a DC value, an edge value, and external information such as a user command or screen content. The encoding may include storing an average or maximum sum of square differences (SSD) value for a detected eye sensitive area and adjusting a QP value based on a comparison of the SSD value to generated threshold values.