Abstract:
An electronic device includes a receiver receives a compressed bitstream and metadata. The electronic device also includes at least one processor that generates an HDR image by decoding the compressed bitstream, identifies viewpoint information based on an orientation of the electronic device, maps the HDR image onto a surface, and renders a portion of the HDR image based on the metadata and the viewpoint information. A display displays the portion of the HDR image.
Abstract:
A method for operating a decoding device for point cloud decoding includes receiving a compressed bitstream. The method also includes decoding the compressed bitstream into two-dimensional (2-D) frames that represent a three-dimensional (3-D) point cloud. Each of the 2-D frames includes a set of patches, and each patch includes a cluster of points of the 3-D point cloud. The cluster of points corresponds to an attribute associated with the 3-D point cloud. One patch of the set of patches, the set of patches, and the 2-D frames correspond to respective access levels representing the 3-D point cloud. The method also includes identifying a first and a second flag. In response to identifying the first and the second flag, the method includes reading the metadata from the bitstream. The method further includes generating, based on metadata and using the sets of 2-D frames, the 3-D point cloud.
Abstract:
A electronic device includes a display and at least one processor. The processor is configured to acquire an image and metadata related to the image. The processor is also configured to identify at least a portion of the image to be enhanced based on the metadata related to the image. The processor is also configured to generate an enhanced image by increasing an intensity of the at least a portion of the image. The processor is also configured to output the enhanced image to the display.
Abstract:
An embodiment of this disclosure provides an audio receiver. The audio receiver includes a memory configured to store an audio signal and processing circuitry coupled to the memory. The processing circuitry is configured to receive the audio signal. The audio signal comprises a plurality of ambisonic components. The processing circuitry is also configured to separate the audio signal into a plurality of independent subcomponents. Each of the independent subcomponents is from a different source. Each of the plurality of ambisonic components is split into the independent subcomponents. The processing circuitry is also configured to decode each of the independent subcomponents. The processing circuitry is also configured to combine each of the decoded independent subcomponents into a speaker signal.
Abstract:
A method for point cloud decoding includes receiving a bitstream. The method also includes decoding the bitstream into multiple frames that include pixels. Certain pixels of the multiple frames correspond to points of a three-dimensional (3D) point cloud. The multiple frames include a first set of frames that represent locations of the points of the 3D point cloud and a second set of frames that represent attribute information for the points of the 3D point cloud. The method further includes reconstructing the 3D point cloud based on the first set of frames. Additionally, the method includes identifying a first portion of the points of the reconstructed 3D point cloud based at least in part on a property associated with the multiple frames. The method also includes modifying a portion of the attribute information. The portion of the attribute information that is modified corresponds to the first portion of the points.
Abstract:
An encoding device and a decoding device is disclosed. The encoding device includes a processor and a communication interface. The processor is configured to generate, for a 3D point cloud, a first frame representing a first attribute and a second frame representing a second attribute. The first and second frames include patches representing respective clusters of points from the 3D point cloud. The processor is configured to generate an occupancy map frame. The processor is configured to identify a querypoint that is positioned on a boundary of one of the patches. The processor is configured to perform smoothing with respect to the querypoint. The processor is configured to encode the frames and generate a compressed bitstream. The communication is configured to transmit the compressed bitstream.
Abstract:
A method for decoding point cloud includes decoding the compressed bitstream into a geometry frame representing geometry of a 3D point cloud and a texture frame representing texture of the 3D point cloud, the geometry frame and the texture frame including pixels representing points of the 3D point cloud from different layers of the 3D point cloud; deriving a set of missing geometry values from pixels in the geometry frame and a set of missing texture values from pixels in the texture frame; generating a first set of frames representing geometry from the different layers based on the geometry frame and the set of missing geometry values; generating a second set of frames representing the texture from the different layers based on the texture frame and the set of missing texture values; and generating the 3D point cloud using the first set of frames and the second set of frames.
Abstract:
A method for processing 360° video is disclosed. The method comprising receiving video data and metadata for processing the video data, the video includes a plurality of regions, determining when the metadata includes an indication that at least one region from the plurality of regions includes a guard margin along at least one boundary of the at least one region, and when the indication indicates that the at least one region includes the guard margin, processing the video data based on the received metadata.
Abstract:
A user equipment (UE) includes a receiver, at least one sensor, and a processor. The receiver is configured to receive a signal comprising at least one encoded image and metadata for the at least one encoded image. The sensor is configured to determine viewpoint information of a user. The processor is configured to render the at least one encoded image based on the metadata and the viewpoint information.
Abstract:
An apparatus includes a memory and a processor. The memory receives a plurality of frames of a scene captured from a camera array. The processor selects a first frame and a second frame from the plurality of frames. The processor also rectifies and aligns the first frame and the second frame to a reference frame, where a blank region of the second frame has a greater area than a blank region of the first frame. The processor further transforms the first frame to have near-optimal superposition to the second frame. The processor inserts a patch from the transformed first frame into the blank region of the second frame.