Symmetric stereo model for handling occlusion
    11.
    发明授权
    Symmetric stereo model for handling occlusion 失效
    用于处理遮挡的对称立体模型

    公开(公告)号:US07599547B2

    公开(公告)日:2009-10-06

    申请号:US11289907

    申请日:2005-11-30

    IPC分类号: G06K9/00 G06T15/00 G01C3/14

    CPC分类号: G06K9/32

    摘要: The present symmetric stereo matching technique provides a method for iteratively estimating a minimum energy for occlusion and disparity using belief propagation. The minimum energy is based on an energy minimization framework in which a visibility constraint is embedded. By embedding the visibility constraint, the present symmetric stereo matching technique treats both images equally, instead of treating one as a reference image. The visibility constraint ensures that occlusion in one view and the disparity in another view are consistent.

    摘要翻译: 本对称立体匹配技术提供了一种使用置信传播迭代估计遮挡和视差的最小能量的方法。 最小能量基于嵌入可见性约束的能量最小化框架。 通过嵌入可见度约束,本对称立体匹配技术可以平等对待两个图像,而不是将其视为参考图像。 可见性约束确保一个视图中的遮挡和另一个视图中的视差是一致的。

    Single-image vignetting correction
    12.
    发明授权
    Single-image vignetting correction 有权
    单图像渐晕校正

    公开(公告)号:US07548661B2

    公开(公告)日:2009-06-16

    申请号:US11384063

    申请日:2006-03-17

    IPC分类号: G06K9/40

    摘要: A system and process for determining the vignetting function of an image and using the function to correct for the vignetting is presented. The image can be any arbitrary image and no other images are required. The system and process is designed to handle both textured and untextured segments in order to maximize the use of available information. To extract vignetting information from an image, segmentation techniques are employed that locate image segments with reliable data for vignetting estimation. Within each image segment, the system and process capitalizes on frequency characteristics and physical properties of vignetting to distinguish it from other sources of intensity variation. The vignetting data acquired from segments are weighted according to a presented reliability measure to promote robustness in estimation.

    摘要翻译: 提出了一种用于确定图像的渐晕功能并使用功能来校正渐晕的系统和过程。 图像可以是任意图像,并且不需要其他图像。 系统和过程被设计为处理纹理和非纹理段,以便最大限度地利用可用信息。 为了从图像中提取渐晕信息,采用定位图像片段以进行晕影估计的可靠数据的分割技术。 在每个图像片段中,系统和过程利用渐晕的频率特性和物理特性来区分其他强度变化的来源。 根据提出的可靠性度量对从片段获取的渐晕数据进行加权,以提高估计的鲁棒性。

    Interactive viewpoint video employing viewpoints forming an array
    14.
    发明授权
    Interactive viewpoint video employing viewpoints forming an array 有权
    使用形成阵列的视点的交互视点视频

    公开(公告)号:US07286143B2

    公开(公告)日:2007-10-23

    申请号:US11097542

    申请日:2005-03-31

    IPC分类号: G09G5/00

    摘要: A system and process for generating, and then rendering and displaying, an interactive viewpoint video in which a user can watch a dynamic scene while manipulating (freezing, slowing down, or reversing) time and changing the viewpoint at will. In general, the interactive viewpoint video is generated using a small number of cameras to capture multiple video streams. A multi-view 3D reconstruction and matting technique is employed to create a layered representation of the video frames that enables both efficient compression and interactive playback of the captured dynamic scene, while at the same time allowing for real-time rendering.

    摘要翻译: 一种用于生成并再现和显示交互视点视频的系统和过程,其中用户可以在操纵(冻结,减慢或反转)时间并随意改变视点的同时观看动态场景。 通常,使用少量摄像机生成交互视点视频以捕获多个视频流。 采用多视图3D重建和消隐技术来创建视频帧的分层表示,使得能够实现捕获的动态场景的高效压缩和交互式回放,同时允许实时渲染。

    System and process for generating a two-layer, 3D representation of a scene
    15.
    发明授权
    System and process for generating a two-layer, 3D representation of a scene 有权
    用于生成场景的两层3D表示的系统和过程

    公开(公告)号:US07206000B2

    公开(公告)日:2007-04-17

    申请号:US11334591

    申请日:2006-01-17

    IPC分类号: G09G5/02

    CPC分类号: G06T15/205

    摘要: A system and process for generating a two-layer, 3D representation of a digital or digitized image from the image and a pixel disparity map of the image is presented. The two layer representation includes a main layer having pixels exhibiting background colors and background disparities associated with correspondingly located pixels of depth discontinuity areas in the image, as well as pixels exhibiting colors and disparities associated with correspondingly located pixels of the image not found in these depth discontinuity areas. The other layer is a boundary layer made up of pixels exhibiting foreground colors, foreground disparities and alpha values associated with the correspondingly located pixels of the depth discontinuity areas. The depth discontinuity areas correspond to prescribed sized areas surrounding depth discontinuities found in the image using a disparity map thereof.

    摘要翻译: 提出了一种用于从图像生成数字或数字化图像的二层3D表示和图像的像素视差图的系统和过程。 两层表示包括具有显示背景颜色的像素和与图像中的深度不连续区域的相应定位的像素相关联的背景差异的主层以及与在这些深度中未找到的图像的相应定位的像素相关联的颜色和差异的像素 不连续区域。 另一层是由与前述深度不连续区域的对应位置的像素相关联的前景色,前景差异和α值的像素构成的边界层。 深度不连续区域对应于使用其视差图在图像中发现的围绕深度不连续性的规定尺寸的区域。

    Compressing and decompressing multiple, layered, video streams employing multi-directional spatial encoding
    16.
    发明授权
    Compressing and decompressing multiple, layered, video streams employing multi-directional spatial encoding 有权
    使用多向空间编码压缩和解压缩多个分层的视频流

    公开(公告)号:US08774274B2

    公开(公告)日:2014-07-08

    申请号:US13348262

    申请日:2012-01-11

    IPC分类号: H04N7/12

    摘要: A process for compressing and decompressing non-keyframes in sequential sets of contemporaneous video frames making up multiple video streams where the video frames in a set depict substantially the same scene from different viewpoints. Each set of contemporaneous video frames has a plurality frames designated as keyframes with the remaining being non-keyframes. In one embodiment, the non-keyframes are compressed using a multi-directional spatial prediction technique. In another embodiment, the non-keyframes of each set of contemporaneous video frames are compressed using a combined chaining and spatial prediction compression technique. The spatial prediction compression technique employed can be a single direction technique where just one reference frame, and so one chain, is used to predict each non-keyframe, or it can be a multi-directional technique where two or more reference frames, and so chains, are used to predict each non-keyframe.

    摘要翻译: 一种用于在构成多个视频流的同步视频帧的顺序集合中压缩和解压缩非关键帧的过程,其中集合中的视频帧从不同视点描绘基本相同的场景。 每组同时期的视频帧具有指定为关键帧的多个帧,其余的是非关键帧。 在一个实施例中,使用多方向空间预测技术来压缩非关键帧。 在另一个实施例中,使用组合链接和空间预测压缩技术来压缩每组同时期视频帧的非关键帧。 所使用的空间预测压缩技术可以是单向技术,其中仅使用一个参考帧,因此使用一条链来预测每个非关键帧,或者它可以是多方向技术,其中两个或更多个参考帧等 链,用于预测每个非关键帧。

    Color segmentation-based stereo 3D reconstruction system and process
    17.
    发明授权
    Color segmentation-based stereo 3D reconstruction system and process 有权
    基于颜色分割的立体3D重建系统和过程

    公开(公告)号:US07324687B2

    公开(公告)日:2008-01-29

    申请号:US10879327

    申请日:2004-06-28

    IPC分类号: G06K9/00

    CPC分类号: G06K9/20 G06K2209/40 G06T7/55

    摘要: A system and process for computing a 3D reconstruction of a scene from multiple images thereof, which is based on a color segmentation-based approach, is presented. First, each image is independently segmented. Second, an initial disparity space distribution (DSD) is computed for each segment, using the assumption that all pixels within a segment have the same disparity. Next, each segment's DSD is refined using neighboring segments and its projection into other images. The assumption that each segment has a single disparity is then relaxed during a disparity smoothing stage. The result is a disparity map for each image, which in turn can be used to compute a per pixel depth map if the reconstruction application calls for it.

    摘要翻译: 提出了一种用于基于基于颜色分割的方法从其多个图像计算场景的3D重建的系统和过程。 首先,每个图像被独立地分割。 第二,使用假设一个段内的所有像素具有相同的视差来为每个段计算初始视差空间分布(DSD)。 接下来,每个段的DSD使用相邻段进行细化,并将其投影到其他图像中。 然后在视差平滑阶段放宽每个段具有单个视差的假设。 结果是每个图像的视差图,如果重建应用程序需要它,它又可以用于计算每像素深度图。

    Real-time rendering system and process for interactive viewpoint video
    18.
    发明授权
    Real-time rendering system and process for interactive viewpoint video 有权
    实时渲染系统和进程的交互视点视频

    公开(公告)号:US07221366B2

    公开(公告)日:2007-05-22

    申请号:US10910088

    申请日:2004-08-03

    IPC分类号: G06T15/20

    CPC分类号: G06T15/205

    摘要: A system and process for rendering and displaying an interactive viewpoint video is presented in which a user can watch a dynamic scene while manipulating (freezing, slowing down, or reversing) time and changing the viewpoint at will. The ability to interactively control viewpoint while watching a video is an exciting new application for image-based rendering. Because any intermediate view can be synthesized at any time, with the potential for space-time manipulation, this type of video has been dubbed interactive viewpoint video.

    摘要翻译: 呈现用于呈现和显示交互式视点视频的系统和过程,其中用户可以在操纵(冻结,减速或反转)时间并随意改变视点的同时观看动态场景。 在观看视频时交互控制视点的能力是基于图像的渲染的令人兴奋的新应用。 因为任何中间视图可以随时被合成,具有时空操纵的潜力,这种类型的视频被称为交互视点视频。

    System and process for generating high dynamic range images from multiple exposures of a moving scene
    19.
    发明授权
    System and process for generating high dynamic range images from multiple exposures of a moving scene 有权
    用于从移动场景的多次曝光中产生高动态范围图像的系统和过程

    公开(公告)号:US07142723B2

    公开(公告)日:2006-11-28

    申请号:US10623033

    申请日:2003-07-18

    IPC分类号: G06K9/40 H04N5/235

    CPC分类号: G06T5/50 G06T7/269

    摘要: A system and process for generating a high dynamic range (HDR) image from a bracketed image sequence, even in the presence of scene or camera motion, is presented. This is accomplished by first selecting one of the images as a reference image. Then, each non-reference image is registered with another one of the images, including the reference image, which exhibits an exposure that is both closer to that of the reference image than the image under consideration and closest among the other images to the exposure of the image under consideration, to generate a flow field. The flow fields generated for the non-reference images not already registered with the reference image are concatenated to register each of them with the reference image. Each non-reference image is then warped using its associated flow field. The reference image and the warped images are combined to create a radiance map representing the HDR image.

    摘要翻译: 提出了即使在存在场景或相机运动的情况下也可以从包围的图像序列生成高动态范围(HDR)图像的系统和过程。 这是通过首先选择一个图像作为参考图像来实现的。 然后,将每个非参考图像与包括参考图像的另一个图像一起登记,该参考图像表现出比正在考虑的图像更接近参考图像的曝光,并且在其他图像中最接近曝光 考虑的图像,产生一个流场。 为未参考图像注册的非参考图像生成的流场被连接以将它们注册到参考图像。 然后使用其相关联的流场对每个非参考图像进行翘曲。 参考图像和翘曲图像被组合以产生表示HDR图像的辐射图。

    System and process for compressing and decompressing multiple, layered, video streams employing spatial and temporal encoding
    20.
    发明授权
    System and process for compressing and decompressing multiple, layered, video streams employing spatial and temporal encoding 有权
    用于使用空间和时间编码来压缩和解压缩多个分层的视频流的系统和过程

    公开(公告)号:US07561620B2

    公开(公告)日:2009-07-14

    申请号:US10910077

    申请日:2004-08-03

    摘要: A system and process for compressing and decompressing multiple video streams depicting substantially the same dynamic scene from different viewpoints. Each frame in each contemporaneous set of video frames of the multiple streams is represented by at least a two layers—a main layer and a boundary layer. Compression of the main layers involves first designating one or more of these layers in each set of contemporaneous frames as keyframes. For each set of contemporaneous frames in time sequence order, the main layer of each keyframe is compressed using an inter-frame compression technique. In addition, the main layer of each non-keyframe within the frame set under consideration is compressed using a spatial prediction compression technique. Finally, the boundary layers of each frame in the current frame set are each compressed using an intra-frame compression technique. Decompression is generally the reverse of the compression process.

    摘要翻译: 一种用于压缩和解压缩从不同观点描绘基本相同的动态场景的多个视频流的系统和过程。 多个流的每个同期视频帧集合中的每个帧由至少两层(主层和边界层)表示。 主要层的压缩包括首先将每组同期帧中的这些层中的一个或多个指定为关键帧。 对于按时间顺序排列的每组同期帧,使用帧间压缩技术对每个关键帧的主层进行压缩。 另外,使用空间预测压缩技术对所考虑的帧集合内的每个非关键帧的主层进行压缩。 最后,使用帧内压缩技术对当前帧集合中每帧的边界层进行压缩。 压缩通常与压缩过程相反。