Abstract:
Coding techniques for image data may cause a still image to be converted to a “phantom” video sequence, which is coded by motion compensated prediction techniques. Thus, coded video data obtained from the coding operation may include temporal prediction references between frames of the video sequence. Metadata may be generated that identifies allocations of content from the still image to the frames of the video sequence. The coded data and the metadata may be transmitted to another device, whereupon it may be decoded by motion compensated prediction techniques and converted back to a still image data. Other techniques may involve coding an image in both a base layer representation and at least one coded enhancement layer representation. The enhancement layer representation may be coded predictively with reference to the base layer representation. The coded base layer representation may be partitioned into a plurality of individually-transmittable segments and stored. Prediction references of elements of the enhancement layer representation may be confined to segments of the base layer representation that correspond to a location of those elements. Meaning, when a pixel block of an enhancement layer maps to a given segment of the base layer representation, prediction references are confined to that segment and do not reference portions of the base layer representation that may be found in other segment(s).
Abstract:
System and methods for improved playback of a video stream are presented. Video snippets are identified that include a number of consecutive frames for playback. Snippets may be evenly temporally spaced in the video stream or may be content adaptive. Then the first frame of a snippet may be selected as the first frame of a scene or other appropriate stopping point. Scene detection, object detection, motion detection, video metadata, or other information generated during encoding or decoding of the video stream may aid in appropriate snippet selection.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding. The encoder is also configured to determine quantized minimum depths and/or maximum depths patch images in the image frames, wherein depth information is signaled relative to the quantized minimum depth.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. A point cloud attribute transfer algorithm may be used to determine distortion between an original point cloud and a reconstructed point cloud. Additionally, the point cloud attribute transfer algorithm may be used to select attribute values for a reconstructed point cloud such that distortion between an original point cloud and a reconstructed version of the original point cloud is minimized.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial information for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding. Various compression strategies may be used to encode an occupancy map and related block-to-patch information indicating which portions of the image frame correspond to which packed patches. Packed image frames comprising patches and padding, along with an encoded occupancy map and related block-to-patch information are sent to a decoder. The decoder is configured to generate a decompressed point cloud based on the packed image frames, compressed occupancy map, and related block-to-patch information.
Abstract:
A system comprises an encoder configured to compress attribute information for a point cloud and/or a decoder configured to decompress compressed attribute for the point cloud. To compress the attribute information, attribute values are predicted using one of a plurality of prediction strategies, wherein a selected prediction strategy is selected based at least in part on attribute variability of points in a neighborhood of points. A decoder follows a similar prediction process. Also, attribute correction values may be determined to correct predicted attribute values and may be used by a decoder to decompress a point cloud, wherein the decoder applies the same prediction strategy applied at the encoder.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. To compress the attribute and/or spatial information, the encoder is configured to convert a point cloud into an image based representation. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud. A processing/filtering element utilizes occupancy map information and/or auxiliary patch information to determine relationships between patches in image frames and adjusts encoding/decoding and/or filtering or pre/post-processing parameters based on the determined relationships.
Abstract:
Video compression and decompression techniques are disclosed that provide improved bandwidth control for video compression and decompression systems. In particular, video coding and decoding techniques quantize input video in multiple dimensions. According to these techniques, pixel residuals may be generated from a comparison of an array of input data to an array of prediction data. The pixel residuals may be quantized in a first dimension. After the quantization, the quantized pixel residuals may be transformed to an array of transform coefficients. The transform coefficients may be quantized in a second dimension and entropy coded. Decoding techniques invert these processes. In still other embodiments, multiple quantizers may be provided upstream of the transform stage, either in parallel or in cascade, which provide greater flexibility to video coders to quantize data in different dimensions in an effort to balance the competing interest in compression efficiency and quality of reconstructed video.
Abstract:
A system comprises an encoder configured to compress attribute information and/or spatial for a point cloud and/or a decoder configured to decompress compressed attribute and/or spatial information for the point cloud. The encoder is configured to convert a point cloud into an image based representation. The encoder packs patch images into an image frame and fills empty spaces in the image frame with a padding, wherein pixel values for the padding are determined based on neighboring pixels values such that the padding is smoothed in the image frame. Also, the decoder is configured to generate a decompressed point cloud based on an image based representation of a point cloud.