AUTOMATIC VIEW MAPPING FOR SINGLE-IMAGE AND MULTI-VIEW CAPTURES

    公开(公告)号:US20200234397A1

    公开(公告)日:2020-07-23

    申请号:US16518501

    申请日:2019-07-22

    Applicant: Fyusion, Inc.

    Abstract: A three-dimensional (3D) skeleton may be determined based on a plurality of vertices and a plurality of faces in a two-dimensional (2D) mesh in a top-down image of an object. A correspondence mapping between a designated perspective view image and the top-down object image may be determined based on the 3D skeleton. The correspondence mapping may link a respective first location in the top-down object image to a respective second location in the designated perspective view image for each of a plurality of points in the designated perspective view image. A top-down mapped image of the object may be created by determining a first respective pixel value for each of the first locations, with each first respective pixel value being determined based on a second respective pixel value for the respective second location linked with the respective first location via the correspondence mapping.

    FREE-VIEWPOINT PHOTOREALISTIC VIEW SYNTHESIS FROM CASUALLY CAPTURED VIDEO

    公开(公告)号:US20200228774A1

    公开(公告)日:2020-07-16

    申请号:US16574639

    申请日:2019-09-18

    Applicant: Fyusion, Inc.

    Abstract: An estimated camera pose may be determined for each of a plurality of single plane images of a designated three-dimensional scene. The sampling density of the single plane images may be below the Nyquist rate. However, the sampling density of the single plane images may be sufficiently high such that the single plane images is sufficiently high such that they may be promoted to multiplane images and used to generate novel viewpoints in a light field reconstruction framework. Scene depth information identifying for each of a respective plurality of pixels in the single plane image a respective depth value may be determined for each single plane image. A respective multiplane image including a respective plurality of depth planes may be determined for each single plane image. Each of the depth planes may include a respective plurality of pixels from the respective single plane image.

    METHOD AND APPARATUS FOR 3-D AUTO TAGGING
    54.
    发明申请

    公开(公告)号:US20200167570A1

    公开(公告)日:2020-05-28

    申请号:US16778981

    申请日:2020-01-31

    Applicant: Fyusion, Inc.

    Abstract: A multi-view interactive digital media representation (MVIDMR) of an object can be generated from live images of an object captured from a camera. Selectable tags can be placed at locations on the object in the MVIDMR. When the selectable tags are selected, media content can be output which shows details of the object at location where the selectable tag is placed. A machine learning algorithm can be used to automatically recognize landmarks on the object in the frames of the MVIDMR and a structure from motion calculation can be used to determine 3-D positions associated with the landmarks. A 3-D skeleton associated with the object can be assembled from the 3-D positions and projected into the frames associated with the MVIDMR. The 3-D skeleton can be used to determine the selectable tag locations in the frames of the MVIDMR of the object.

    Client-server communication for live search using multi-view digital media representations

    公开(公告)号:US10353946B2

    公开(公告)日:2019-07-16

    申请号:US15409497

    申请日:2017-01-18

    Applicant: Fyusion, Inc.

    Abstract: Provided are mechanisms and processes for performing live search using multi-view digital media representations. In one example, a process includes receiving a visual search query from a device for an object to be searched, where the visual search query includes a first set of viewpoints of the object obtained during capture of a first surround view of the object during a live search session. Next, additional recommended viewpoints of the object are identified for the device to capture, where the additional recommended viewpoints are chosen to provide more information about the object. A first set of search results based on the first set of viewpoints and additional recommended viewpoints of the object are transmitted to the device. In response, a second set of viewpoints of the object captured using image capture capabilities of the device are received. A second set of search results with enhanced matches for the object based on the first and second sets of viewpoints are then transmitted to the device. This process may continue iteratively until a desired set of search results is obtained.

    CLIENT-SERVER COMMUNICATION FOR LIVE SEARCH USING MULTI-VIEW DIGITAL MEDIA REPRESENTATIONS

    公开(公告)号:US20180203880A1

    公开(公告)日:2018-07-19

    申请号:US15409497

    申请日:2017-01-18

    Applicant: Fyusion, Inc.

    Abstract: Provided are mechanisms and processes for performing live search using multi-view digital media representations. In one example, a process includes receiving a visual search query from a device for an object to be searched, where the visual search query includes a first set of viewpoints of the object obtained during capture of a first surround view of the object during a live search session. Next, additional recommended viewpoints of the object are identified for the device to capture, where the additional recommended viewpoints are chosen to provide more information about the object. A first set of search results based on the first set of viewpoints and additional recommended viewpoints of the object are transmitted to the device. In response, a second set of viewpoints of the object captured using image capture capabilities of the device are received. A second set of search results with enhanced matches for the object based on the first and second sets of viewpoints are then transmitted to the device. This process may continue iteratively until a desired set of search results is obtained.

    REAL-TIME MOBILE DEVICE CAPTURE AND GENERATION OF AR/VR CONTENT

    公开(公告)号:US20170148223A1

    公开(公告)日:2017-05-25

    申请号:US15428104

    申请日:2017-02-08

    Applicant: Fyusion, Inc.

    Abstract: Various embodiments describe systems and processes for generating AR/VR content. In one aspect, a method for generating a three-dimensional (3D) projection of an object is provided. A sequence of images along a camera translation may be obtained using a single lens camera. Each image contains at least a portion of overlapping subject matter, which includes the object. The object is semantically segmented from the sequence of images using a trained neural network to form a sequence of segmented object images, which are then refined using fine-grained segmentation. On-the-fly interpolation parameters are computed and stereoscopic pairs are generated for points along the camera translation from the refined sequence of segmented object images for displaying the object as a 3D projection in a virtual reality or augmented reality environment. Segmented image indices are then mapped to a rotation range for display in the virtual reality or augmented reality environment.

Patent Agency Ranking