SYSTEM AND METHOD FOR SCENE RECONSTRUCTION WITH PLANE AND SURFACE RECONSTRUCTION

    公开(公告)号:US20230088963A1

    公开(公告)日:2023-03-23

    申请号:US17696746

    申请日:2022-03-16

    Abstract: A system and method for 3D reconstruction with plane and surface reconstruction, scene parsing, depth reconstruction with depth fusion from different sources. The system includes display and a processor to perform the method for 3D reconstruction with plane and surface reconstruction. The method includes dividing a scene of an image frame into one or more plane regions and one or more surface regions. The method also includes generating reconstructed planes by performing plane reconstruction based on the one or more plane regions. The method also includes generating reconstructed surfaces by performing surface reconstruction based on the one or more surface regions. The method further includes creating the 3D scene reconstruction by integrating the reconstructed planes and the reconstructed surfaces.

    Method and system for video transformation for video see-through augmented reality

    公开(公告)号:US12154219B2

    公开(公告)日:2024-11-26

    申请号:US18052827

    申请日:2022-11-04

    Abstract: A method of video transformation for a video see-through (VST) augmented reality (AR) device includes obtaining video frames from multiple cameras associated with the VST AR device, where each video frame is associated with position data. The method also includes generating camera viewpoint depth maps associated with the video frames based on the video frames and the position data. The method further includes performing depth re-projection to transform the video frames from camera viewpoints to rendering viewpoints using the camera viewpoint depth maps. The method also includes performing hole filling of one or more holes created in one or more occlusion areas of at least one of the transformed video frames during the depth re-projection to generate at least one hole-filled video frame. In addition, the method includes displaying the transformed video frames including the at least one hole-filled video frame on multiple displays associated with the VST AR device.

    DYNAMIC ALIGNMENT BETWEEN SEE-THROUGH CAMERAS AND EYE VIEWPOINTS IN VIDEO SEE-THROUGH (VST) EXTENDED REALITY (XR)

    公开(公告)号:US20240346779A1

    公开(公告)日:2024-10-17

    申请号:US18630767

    申请日:2024-04-09

    CPC classification number: G06T19/006 G06T5/80 H04N13/344

    Abstract: A method includes determining that an inter-pupillary distance (IPD) between display lenses of a video see-through (VST) extended reality (XR) device has been adjusted with respect to a default IPD. The method also includes obtaining an image captured using a see-through camera of the VST XR device. The see-through camera is configured to capture images of a three-dimensional (3D) scene. The method further includes transforming the image to match a viewpoint of a corresponding one of the display lenses according to a change in IPD with respect to the default IPD in order to generate a transformed image. The method also includes correcting distortions in the transformed image based on one or more lens distortion coefficients corresponding to the change in IPD in order to generate a corrected image. In addition, the method includes initiating presentation of the corrected image on a display panel of the VST XR device.

    DEPTH-VARYING REPROJECTION PASSTHROUGH IN VIDEO SEE-THROUGH (VST) EXTENDED REALITY (XR)

    公开(公告)号:US20240223742A1

    公开(公告)日:2024-07-04

    申请号:US18526726

    申请日:2023-12-01

    CPC classification number: H04N13/344 G06T19/006 H04N13/128 H04N13/239

    Abstract: A method includes obtaining images of a scene captured using a stereo pair of imaging sensors of an XR device and depth data associated with the images, where the scene includes multiple objects. The method also includes obtaining volume-based 3D models of the objects. The method further includes, for one or more first objects, performing depth-based reprojection of the one or more 3D models of the one or more first objects to left and right virtual views based on one or more depths of the one or more first objects. The method also includes, for one or more second objects, performing constant-depth reprojection of the one or more 3D models of the one or more second objects to the left and right virtual views based on a specified depth. In addition, the method includes rendering the left and right virtual views for presentation by the XR device.

    MASK GENERATION WITH OBJECT AND SCENE SEGMENTATION FOR PASSTHROUGH EXTENDED REALITY (XR)

    公开(公告)号:US20240223739A1

    公开(公告)日:2024-07-04

    申请号:US18360677

    申请日:2023-07-27

    CPC classification number: H04N13/128 G06T19/006 H04N2013/0092

    Abstract: A method includes obtaining first and second image frames of a scene. The method also includes providing the first image frame as input to an object segmentation model, where the object segmentation model is trained to generate first object segmentation predictions for objects in the scene and a depth or disparity map based on the first image frame. The method further includes generating second object segmentation predictions for the objects in the scene based on the second image frame. The method also includes determining boundaries of the objects in the scene based on the first and second object segmentation predictions. In addition, the method includes generating a virtual view for presentation on a display of an extended reality (XR) device based on the boundaries of the objects in the scene.

    METHOD AND DEVICE FOR DIRECT PASSTHROUGH IN VIDEO SEE-THROUGH (VST) AUGMENTED REALITY (AR)

    公开(公告)号:US20240062483A1

    公开(公告)日:2024-02-22

    申请号:US18296095

    申请日:2023-04-05

    Inventor: Yingen Xiong

    CPC classification number: G06T19/006 G06V10/761 H04N13/117 H04N13/383

    Abstract: A method includes receiving first and second images from first and second see-through cameras with first and second camera viewpoints. The method also includes generating a first virtual image corresponding to a first virtual viewpoint by applying a first mapping to the first image. The first mapping is based on relative positions of the first camera viewpoint and the first virtual viewpoint corresponding to a first eye of a user. The method further includes generating a second virtual image corresponding to a second virtual viewpoint by applying a second mapping to the second image. The second mapping is based on relative positions of the second camera viewpoint and the second virtual viewpoint corresponding to a second eye of the user. In addition, the method includes presenting the first and second virtual images to the first and second virtual viewpoints on at least one display panel of an augmented reality device.

Patent Agency Ranking