Determining Disparity Search Range in Stereo Videos
    21.
    发明申请
    Determining Disparity Search Range in Stereo Videos 有权
    确定立体声视频中的差异搜索范围

    公开(公告)号:US20110158528A1

    公开(公告)日:2011-06-30

    申请号:US12651200

    申请日:2009-12-31

    IPC分类号: G06T7/00 G06K9/68

    CPC分类号: G06T7/593 G06T2207/10021

    摘要: Embodiments of the invention disclose a system and a method for determining a disparity search range for a current stereo image of a scene based on a set of stereo images of the scene, comprising steps of: selecting a subset of stereo images from the set of stereo images, the subset includes the current stereo image and at least one neighboring stereo image, wherein the neighboring stereo image is temporally-neighboring to the current stereo image; determining a disparity histogram for each stereo image in the subset of stereo images to form a set of disparity histograms; determining a weighted disparity histogram as a weighted sum of the disparity histograms in the set of disparity histograms; and determining the disparity search range from the weighted disparity histogram.

    摘要翻译: 本发明的实施例公开了一种用于基于场景的一组立体图像来确定场景的当前立体图像的视差搜索范围的系统和方法,包括以下步骤:从立体声集合中选择立体图像的子集 图像,所述子集包括当前立体声图像和至少一个邻近立体图像,其中相邻立体图像与当前立体图像在时间上相邻; 确定立体图像子集中的每个立体图像的视差直方图以形成一组视差直方图; 将加权视差直方图确定为视差直方图组中的视差直方图的加权和; 并从加权视差直方图确定视差搜索范围。

    Method and system for randomly accessing multiview videos with known prediction dependency
    22.
    发明授权
    Method and system for randomly accessing multiview videos with known prediction dependency 有权
    随机访问具有已知预测依赖关系的多视点视频的方法和系统

    公开(公告)号:US07903737B2

    公开(公告)日:2011-03-08

    申请号:US11385620

    申请日:2006-03-21

    摘要: A method randomly accesses multiview videos. Multiview videos are acquired of a scene with corresponding cameras arranged at poses, such that there is view overlap between any pair of cameras. V-frames are generated from the multiview videos. The V-frames are encoded using only spatial prediction. Then, the V-frames are inserted periodically in an encoded bit stream to provide random temporal access to the multiview videos. Additional view dependency information enables the decoding of a reduced number of frames prior to accessing randomly a target frame for a specified view and time, and decoding the target frame.

    摘要翻译: 一种随机访问多视点视频的方法。 多视点视频被采集到具有以姿势布置的相应摄像机的场景,使得在任何一对摄像机之间存在视图重叠。 V帧是从多视角视频生成的。 V帧仅使用空间预测编码。 然后,将V帧周期性地插入到编码比特流中,以便为多视点视频提供随机的时间访问。 附加的视图依赖性信息使得能够在对指定的视图和时间的随机访问目标帧之前对减少数量的帧进行解码,以及对目标帧进行解码。

    Method Coding Multi-Layered Depth Images
    23.
    发明申请
    Method Coding Multi-Layered Depth Images 审中-公开
    方法编码多层深度图像

    公开(公告)号:US20100278232A1

    公开(公告)日:2010-11-04

    申请号:US12435057

    申请日:2009-05-04

    IPC分类号: H04N7/26 G06K9/36

    摘要: A method reconstructs a depth image encoded as a base layer bitstream, and a set of enhancement layer bitstreams. The base layer bitstream is decoded to produce pixels of a reconstructed base layer image corresponding to the depth image. Each enhancement layer bitstream is decoded in a low to high order to produces a reconstructed residual image. During the decoding of the enhancement layer bitstream, a context model is maintained using an edge map, and each enhancement layer bitstream is entropy decoded using the context model to determine a significance value corresponding to pixels of the reconstructed residual image and a sign bit for each significant pixel, and a pixel value of the reconstructed residual image is reconstructed according to the significance value, sign bit and an uncertainty interval. Then, the reconstructed residual images are added to the reconstructed base layer image to produce the reconstructed depth image.

    摘要翻译: 一种方法重建被编码为基本层比特流的深度图像,以及一组增强层比特流。 基层比特流被解码以产生对应于深度图像的重建的基本层图像的像素。 每个增强层比特流以低到高的顺序被解码以产生重构的残差图像。 在增强层比特流的解码期间,使用边缘图维持上下文模型,并且使用上下文模型对每个增强层比特流进行熵解码,以确定对应于重构残差图像的像素的显着值和每个 根据显着性值,符号位和不确定性间隔重构重建残差图像的像素值。 然后,将重构的残留图像添加到重构的基本层图像中以产生重建的深度图像。

    Method and system for processing multiview videos for view synthesis using skip and direct modes
    24.
    发明授权
    Method and system for processing multiview videos for view synthesis using skip and direct modes 有权
    用于使用跳过和直接模式处理视频合成的多视点视频的方法和系统

    公开(公告)号:US07671894B2

    公开(公告)日:2010-03-02

    申请号:US11621400

    申请日:2007-01-09

    IPC分类号: H04N5/225

    摘要: A method processes a multiview videos of a scene, in which each video is acquired by a corresponding camera arranged at a particular pose, and in which a view of each camera overlaps with the view of at least one other camera. Side information for synthesizing a particular view of the multiview video is obtained in either an encoder or decoder. A synthesized multiview video is synthesized from the multiview videos and the side information. A reference picture list is maintained for each current frame of each of the multiview videos, the reference picture indexes temporal reference pictures and spatial reference pictures of the acquired multiview videos and the synthesized reference pictures of the synthesized multiview video. Each current frame of the multiview videos is predicted according to reference pictures indexed by the associated reference picture list with a skip mode and a direct mode, whereby the side information is inferred from the synthesized reference picture.

    摘要翻译: 一种处理场景的多视点视频的方法,其中通过布置在特定姿态的相应摄像机获取每个视频,并且其中每个摄像机的视图与至少一个其他摄像机的视图重叠。 在编码器或解码器中获得用于合成多视图视频的特定视图的侧面信息。 合成的多视点视频是从多视点视频和侧面信息合成的。 维护每个多视角视频的每个当前帧的参考图片列表,参考图片索引所获取的多视角视频的时间参考图片和空间参考图片以及合成的多视角视频的合成参考图片。 根据由相​​关联的参考图像列表以跳过模式和直接模式索引的参考图像来预测多视图视频的每个当前帧,由此从合成的参考图片推断出侧面信息。

    Filtering Artifacts in Images with 3D Spatio-Temporal Fuzzy Filters
    25.
    发明申请
    Filtering Artifacts in Images with 3D Spatio-Temporal Fuzzy Filters 失效
    用3D空间时间模糊滤镜过滤人像

    公开(公告)号:US20080019605A1

    公开(公告)日:2008-01-24

    申请号:US11866476

    申请日:2007-10-03

    IPC分类号: G06K9/36 G06K9/00

    摘要: A method filters pixels in a sequence of images. Each image in the sequence is partitioned into blocks of pixels, and the images are processed sequentially. The energy is determined for each block of pixels in each image. The energy of each block is based on variances of intensities of the pixels in the sequence of images. A 3D fuzzy filter is applied to each current pixel in each current block during the sequential processing. The 3D fuzzy filter considers the energy of the block, and the intensities of pixels spatially adjacent and temporally adjacent to the current pixel to remove blocking and ringing artifacts.

    摘要翻译: 一种方法可以过滤图像序列中的像素。 序列中的每个图像被分割成像素块,并且顺序地处理图像。 为每个图像中的每个像素块确定能量。 每个块的能量基于图像序列中的像素的强度的方差。 在顺序处理期间,将3D模糊滤波器应用于每个当前块中的每个当前像素。 3D模糊滤波器考虑了块的能量,以及与当前像素空间相邻和时间相邻的像素的强度,以消除阻塞和振铃伪像。

    Method for enhancing depth images of scenes using trellis structures
    26.
    发明授权
    Method for enhancing depth images of scenes using trellis structures 有权
    使用网格结构增强场景深度图像的方法

    公开(公告)号:US08994722B2

    公开(公告)日:2015-03-31

    申请号:US13406139

    申请日:2012-02-27

    IPC分类号: G06T15/20

    CPC分类号: G06T15/205

    摘要: An image for a virtual view of a scene is generated based on a set of texture images and a corresponding set of depth images acquired of the scene. A set of candidate depths associated with each pixel of a selected image is determined. For each candidate depth, a cost that estimates a synthesis quality of the virtual image is determined. The candidate depth with a least cost is selected to produce an optimal depth for the pixel. Then, the virtual image is synthesized based on the optimal depth of each pixel and the texture images. The method also applies first and second depth enhancement before, and during view synthesis to correct errors or suppress noise due to the estimation or acquisition of the dense depth images and sparse depth features.

    摘要翻译: 基于一组纹理图像和从该场景获取的对应的一组深度图像来生成用于场景的虚拟视图的图像。 确定与所选图像的每个像素相关联的一组候选深度。 对于每个候选深度,确定估计虚拟图像的合成质量的成本。 选择具有最低成本的候选深度以产生像素的最佳深度。 然后,基于每个像素的最佳深度和纹理图像合成虚拟图像。 该方法还在视图合成之前和视图合成期间都应用第一和第二深度增强,以校正误差或抑制由于估计或获取密集深度图像和稀疏深度特征而导致的噪声。

    Method for selecting transform types from mapping table for prediction modes
    27.
    发明授权
    Method for selecting transform types from mapping table for prediction modes 有权
    从预测模式的映射表中选择变换类型的方法

    公开(公告)号:US08929455B2

    公开(公告)日:2015-01-06

    申请号:US13175737

    申请日:2011-07-01

    摘要: A method codes pictures in a bitstream, wherein the bitstream includes coded pictures to obtain data for associated TUs and data for generating a transform tree, and a partitioning of coding units (CUs) into Prediction Units (PUs), and data for obtaining prediction modes or directions associated with each PU. One or more mapping tables are defined, wherein each row of each table has an associated index and a first set of transform types to be used for applying an inverse transformation to the data in TU. The first set of transform types is selected according to an index, and then a second set of transform types is applied as the inverse transformation to the data, wherein the second set of transform types is determined according to the first set of transform types and a transform-toggle flag (ttf) to obtain a reconstructed prediction residual.

    摘要翻译: 一种方法对比特流中的图像进行编码,其中比特流包括用于获得相关联的TU的数据的编码图像和用于生成变换树的数据,以及将编码单元(CU)划分为预测单元(PU)和用于获得预测模式的数据 或与每个PU相关联的方向。 定义一个或多个映射表,其中每个表的每一行具有相关联的索引和用于将逆变换应用于TU中的数据的第一组变换类型。 根据索引选择第一组变换类型,然后应用第二组变换类型作为数据的逆变换,其中根据第一组变换类型确定第二组变换类型,并且a 变换切换标志(ttf)以获得重建的预测残差。

    Method for coding videos using dictionaries
    28.
    发明授权
    Method for coding videos using dictionaries 有权
    使用字典编码视频的方法

    公开(公告)号:US08767835B2

    公开(公告)日:2014-07-01

    申请号:US12979414

    申请日:2010-12-28

    摘要: A video encoded as a bit stream is decoded by maintaining a set of dictionaries generated from decoded prediction residual signals, wherein elements of the set of dictionaries have associated indices. A current macroblock is entropy decoded and inverse quantized to produce decoded coefficients. For the current macroblock, a particular dictionary of the set of dictionaries is selected according to a prediction mode signaled in the bit stream, and particular elements of the particular dictionary are selected according to a copy mode signal in the bit stream and the associated index. The particular elements is scaled and combined, using the decoded coefficients, to reconstruct a current decoded macroblock prediction residual signal. Then, the current decoded macroblock prediction residual signal is combined with previously decoded macroblocks to generate an output macroblock of a reconstructed video, wherein the steps are performed in a decoder.

    摘要翻译: 通过维持从解码的预测残差信号产生的一组字典来解码编码为比特流的视频,其中字典集合中的元素具有相关联的索引。 当前宏块被熵解码并被反量化以产生解码系数。 对于当前的宏块,根据在位流中发信号的预测模式来选择字典集合的特定字典,并且根据位流中的复制模式信号和相关联的索引来选择特定字典的特定元素。 使用解码的系数对特定元素进行缩放和组合,以重构当前解码的宏块预测残差信号。 然后,将当前解码的宏块预测残差信号与先前解码的宏块相组合以产生重构视频的输出宏块,其中在解码器中执行这些步骤。

    Method for coding pictures using hierarchical transform units
    29.
    发明授权
    Method for coding pictures using hierarchical transform units 有权
    使用分层变换单元编码图像的方法

    公开(公告)号:US08494290B2

    公开(公告)日:2013-07-23

    申请号:US13169959

    申请日:2011-06-27

    IPC分类号: G06K9/36

    摘要: A bitstream includes coded pictures, and split-flags for generating a transform tree. The bit stream is a partitioning of coding units (CUs) into Prediction Units (PUs). The transform tree is generated according to the split-flags. Nodes in the transform tree represent transform units (TU) associated with the CUs. The generation splits each TU only if the corresponding split-flag is set. For each PU that includes multiple TUs, the multiple TUs are merged into a larger TU, and the transform tree is modified according to the splitting and merging. Then, data contained in each PU can be decoded using the TUs associated with the PU according to the transform tree.

    摘要翻译: 比特流包括编码图像,以及用于生成变换树的分割标志。 比特流是将编码单元(CU)划分成预测单元(PU)。 转换树根据分裂标志生成。 变换树中的节点表示与CU相关联的变换单元(TU)。 仅当相应的分割标志被设置时,该代码才能分割每个TU。 对于包含多个TU的每个PU,将多个TU合并成一个较大的TU,根据分割和合并修改变换树。 然后,可以使用根据变换树与PU相关联的TU来解码包含在每个PU中的数据。

    Method for Distributed Source Coding of Wavelet Coefficients in Zerotrees
    30.
    发明申请
    Method for Distributed Source Coding of Wavelet Coefficients in Zerotrees 有权
    Zerotrees中小波系数的分布式源码编码方法

    公开(公告)号:US20130051691A1

    公开(公告)日:2013-02-28

    申请号:US13218125

    申请日:2011-08-25

    IPC分类号: G06K9/36

    摘要: A method for encoding a source image, wherein the source image includes a set of bitplanes of pixels, is disclosed. For each bitplane in a most to least significant order, the method include obtaining a list of significant pixels (LSP), a list of insignificant pixels (LIP), and a list of insignificant sets (LIS) according to a hierarchical ordering of the source image pixels; synchronizing the LSP, LIP and LIS of the source image with the LSP, LIP and LIS of a key image; constructing a temporary list of insignificant sets (TLIS) for the source image; and applying syndrome encoding to the LSP, LIP, and TLIS of the source image to obtain syndromes corresponding to magnitudes and signs of pixels in the source image, wherein the steps are performed in a processor.

    摘要翻译: 公开了一种用于对源图像进行编码的方法,其中源图像包括一组像素位平面。 对于每个位平面以最低至最低有序的顺序,该方法包括根据源的分层顺序获得有效象素列表(LSP),无效像素列表(LIP)和无效集列表(LIS) 图像像素; 将源图像的LSP,LIP和LIS与关键图像的LSP,LIP和LIS进行同步; 为源图像构建一个不重要的集合(TLIS)的临时列表; 以及对源图像的LSP,LIP和TLIS应用校正子编码,以获得与源图像中的像素的大小和符号相对应的校正子,其中步骤在处理器中执行。