-
公开(公告)号:WO2023278809A1
公开(公告)日:2023-01-05
申请号:PCT/US2022/035881
申请日:2022-06-30
Applicant: ZOOM VIDEO COMMUNICATIONS, INC.
Inventor: BALAJI, Abhishek , LING, Bo , PARK, Juliana , WALIA, Nitasha , WANG, Jianpeng , WANG, Ruizhen
IPC: G06T5/00 , G06T7/11 , G06T7/73 , G06T7/77 , G06T2207/10024 , G06T2207/20028 , G06T2207/20132 , G06T2207/30201 , G06T5/002 , G06T7/74
Abstract: Methods and systems provide for video appearance adjustments within a video communication session. First, the system receives video content. The system then receives an appearance adjustment request comprising an adjustment depth, and detects imagery of a user within the video content. The system then detects a face region within the video content. The system segments the face region into a number of skin areas. For each of the plurality of skin areas, the system classifies the skin area as a smooth texture region or rough texture region. If the skin area is classified as a smooth texture region, the system modifies the imagery of the user in real time or substantially real time by applying a smoothing process to the skin area, where the amount of smoothing applied corresponds to the adjustment depth.
-
公开(公告)号:WO2022207687A2
公开(公告)日:2022-10-06
申请号:PCT/EP2022/058383
申请日:2022-03-30
Applicant: XYZ REALITY LIMITED
Inventor: AHMED, Umar , MITCHELL, David
IPC: G01C15/00 , G01C15/06 , G06F3/01 , G01C15/002 , G02B27/01 , G06F3/012 , G06F3/013 , G06T19/006 , G06T7/248 , G06T7/74
Abstract: Certain examples described herein present a method of configuring a headset for display of a building information model (BUM) at a construction site. In one example, the method includes: determining a pose of the headset using a positioning system; obtaining an image of a two-dimensional marker positioned within the construction site; determining a BIM-to-camera transformation between an origin of a coordinate system used by the building information model and an origin of the coordinate system of the camera based on a location of the two-dimensional marker within the image; determining a camera-to-positioning transformation between the origin of the coordinate system used by the camera and an origin of a coordinate system used by the positioning system; and determining a BIM-to-positioning transformation.
-
公开(公告)号:WO2021202785A1
公开(公告)日:2021-10-07
申请号:PCT/US2021/025237
申请日:2021-03-31
Applicant: WOVEN PLANET NORTH AMERICA, INC.
Inventor: RICHERT, Wilhelm , SCHIWIETZ, Thomas
Abstract: Examples disclosed herein may involve (i) obtaining an aggregated overhead view image of a geographical area that has been generated by a pipeline for generating aggregated overhead view images, the geographical area comprising a plurality of regions, where the aggregated overhead view image is generated from aggregating pixel values from a plurality of source images of the geographical area, (ii) generating one or more reprojection images of one or more of the regions of the geographic area from the aggregated overhead view image, (iii) identifying, from the plurality of source images, one or more source images that capture the one or more regions of the geographical area, (iv) calculating one or more differences between the identified one or more source images and the one or more reprojection images, and (v) determining one or more error corrections to be applied to the pipeline for generating overhead view images
-
公开(公告)号:WO2021155246A1
公开(公告)日:2021-08-05
申请号:PCT/US2021/015850
申请日:2021-01-29
Applicant: HOVER INC.
Inventor: CASTILLO, William , SCOTT, Brandon , FIRL, Alrik , CUTTS, David Royston , IGNER, Jonathan Mark , CURRO, Domenico , RETHAGE, Dario , LI, Panfeng
IPC: H04N5/232 , G06F3/167 , G06T15/00 , G06T17/00 , G06T2207/20072 , G06T2207/20084 , G06T2210/00 , G06T7/11 , G06T7/12 , G06T7/174 , G06T7/277 , G06T7/74 , G06V10/26 , G06V10/44 , G06V30/19013 , G06V30/19107 , G06V30/414 , H04N5/23218 , H04N5/23222 , H04N5/232945
Abstract: Disclosed are techniques for enhancing two-dimensional (2D) image capture of subjects (e.g., a physical structure, such as a residential building) to maximize the feature correspondences available for three-dimensional (3D) model reconstruction. More specifically, disclosed is a computer-vision network configured to provide viewfinder interfaces and analyses to guide the improved capture of an intended subject for specified purposes. Additionally, the computer-vision network can be configured to generate a metric representing a quality of feature correspondences between images of a complete set of images used for reconstructing a 3D model of a physical structure. The computer-vision network can also be configured to generate feedback at or before image capture time to guide improvements to the quality of feature correspondences between a pair of images.
-
公开(公告)号:WO2021145909A1
公开(公告)日:2021-07-22
申请号:PCT/US2020/026954
申请日:2020-04-06
Applicant: RAYTHEON COMPANY
Inventor: NGUYEN, Huy, P. , KRAUSSER, Dieter, G. , KANNAN, Pradyumna
IPC: G06T7/73 , G01C21/02 , G01C21/025 , G06K9/0063 , G06K9/4609 , G06K9/6202 , G06T2207/10032 , G06T2207/30244 , G06T7/248 , G06T7/74 , G06T7/75
Abstract: A system and method for star tracking includes: capturing an image of stars; detecting and selecting visible stars from the captured image; extracting features from the selected stars by forming a convex hull from the selected stars to generate a spherical polygon; computing the area and higher order moments of the spherical polygon; and pattern matching the extracted feature against a database of star catalog. The pattern matching includes matching the area of the spherical polygon to a plurality of polygon areas stored in the database and when the number of the matching candidates is more than one, matching a next extracted higher order moment with a respective higher order moment in the database, and repeating said matching of the next extracted higher order moment until the number of the matching candidates is equal to one.
-
公开(公告)号:WO2023278085A1
公开(公告)日:2023-01-05
申请号:PCT/US2022/031664
申请日:2022-05-31
Applicant: MICROSOFT TECHNOLOGY LICENSING, LLC
Inventor: PUTTASWAMY, Niranjan Achugundla , GIBSON, Gregory Theodore , MARGOLIS, Jeffrey Neil , TARDIF, John Allen
IPC: G02B27/00 , H04N5/347 , H04N5/359 , G06F3/01 , A61B3/14 , G02B27/0093 , G02B5/20 , G06F3/013 , G06T2207/30201 , G06T5/20 , G06T7/248 , G06T7/74 , G06V40/19 , G06V40/197 , H04N25/46 , H04N25/62 , H04N25/75
Abstract: Technologies are described herein for an eye tracking that may be employed by devices and systems such as head mount display (HMD) devices. Light that is reflected from a user's eye may be specular or scattered. The specular light has an intensity or magnitude that may saturate the electronics. The presently disclosed techniques mitigate saturation by generating detected signals from an optical detector, evaluating the signal levels for the detected signal, and selectively gating the detected signals that have saturated. The remaining scattered signals can be combined to achieve a combined signal that can be converted into a digital signal without saturating the electronics, which can then be processed to form an image of the eye for identification purposes, for tracking eye movement, and for other uses. The described technologies provide a clear image without ambient light reflections or specular light interfering with the image.
-
公开(公告)号:WO2022192266A2
公开(公告)日:2022-09-15
申请号:PCT/US2022/019354
申请日:2022-03-08
Applicant: RIDECELL, INC.
IPC: G06T7/11 , B60W2420/52 , B60W2554/4029 , B60W2554/4041 , B60W60/0015 , G01S17/42 , G01S17/89 , G01S17/931 , G06N5/027 , G06T2207/10016 , G06T2207/10028 , G06T2207/20081 , G06T2207/20084 , G06T2207/30196 , G06T2207/30252 , G06T2207/30261 , G06T7/521 , G06T7/73 , G06T7/74
Abstract: Multi-object tracking in autonomous vehicles uses both camera data and LiDAR data for training, but not LiDAR data at query time. Thus, no LiDAR sensor is on a piloted autonomous vehicle. Example systems and methods rely on camera 2D object detections alone, rather than 3D annotations. Example systems/methods utilize a single network that is given a camera image as input and can learn both object detection and dense depth in a multimodal regression setting, where the ground truth LiDAR data is used only at training time to compute depth regression loss. The network uses the camera image alone as input at test time (i.e., when deployed for piloting an autonomous vehicle) and can predict both object detections and dense depth of the scene. LiDAR is only used for data acquisition and is not required for drawing 3D annotations or for piloting the vehicle.
-
公开(公告)号:WO2021262943A1
公开(公告)日:2021-12-30
申请号:PCT/US2021/038835
申请日:2021-06-24
Applicant: AMAZON TECHNOLOGIES, INC.
Inventor: BOYRAZ, Hakan , LIU, Baoyuan , NIE, Xiaohan , CHEN, Sheng
IPC: G06K9/00 , G06K9/62 , B60W60/00256 , G05D1/0094 , G05D1/024 , G05D1/0274 , G06K9/6288 , G06T2207/10024 , G06T2207/10028 , G06T2207/30261 , G06T7/74 , G06V20/10 , G06V20/58 , G06V20/64
Abstract: The present disclosure generally relates to a system of a delivery device for combining sensor data from various types of sensors to generate a map that enables the delivery device to navigate from a first location to a second location to deliver an item to the second location. The system obtains data from RGB, LIDAR, and depth sensors and combines this sensor data according to various algorithms to detect objects in an environment of the delivery device, generate point cloud and pose information associated with the detected objects, and generates object boundary data for the detected objects. The system further identifies object states for the detected object and generates the map for the environment based on the detected object, the generated object proposal data, the labeled point cloud data, and the object states. The generated map may be provided to other systems to navigate the delivery device.
-
公开(公告)号:WO2021155246A
公开(公告)日:2021-08-05
申请号:PCT/US2021/015850
申请日:2021-01-29
Applicant: HOVER INC. [US]/[US]
Inventor: CASTILLO, William , SCOTT, Brandon , FIRL, Alrik , CUTTS, David, Royston , IGNER, Jonathan, Mark , CURRO, Domenico , RETHAGE, Dario , LI, Panfeng
IPC: H04N5/232 , G06F3/167 , G06T17/00 , G06T2207/20072 , G06T2207/20084 , G06T7/11 , G06T7/12 , G06T7/174 , G06T7/74 , H04N5/23218 , H04N5/23222 , H04N5/232945
Abstract: Disclosed are techniques for enhancing two-dimensional (2D) image capture of subjects (e.g., a physical structure, such as a residential building) to maximize the feature correspondences available for three-dimensional (3D) model reconstruction. More specifically, disclosed is a computer-vision network configured to provide viewfinder interfaces and analyses to guide the improved capture of an intended subject for specified purposes. Additionally, the computer-vision network can be configured to generate a metric representing a quality of feature correspondences between images of a complete set of images used for reconstructing a 3D model of a physical structure. The computer-vision network can also be configured to generate feedback at or before image capture time to guide improvements to the quality of feature correspondences between a pair of images.
-
10.
公开(公告)号:WO2021146118A1
公开(公告)日:2021-07-22
申请号:PCT/US2021/012777
申请日:2021-01-08
Applicant: TRIMBLE INC.
Inventor: NICHOLS, Mark
IPC: G06F3/01 , G06T19/00 , G06T7/70 , B60K35/00 , G06F3/03 , E02F3/00 , B60K2370/1529 , B60K2370/177 , B60K2370/61 , E02F9/205 , E02F9/261 , G02B2027/0181 , G02B2027/0183 , G02B2027/0187 , G02B27/0172 , G02B27/0179 , G05D1/0253 , G05D1/027 , G05D1/0278 , G06F3/011 , G06F3/012 , G06F3/0304 , G06K9/00671 , G06T19/006 , G06T2207/30204 , G06T2207/30244 , G06T7/74
Abstract: Systems and methods described herein provide augmented reality images to an operator of a machine. A pose of an augmented reality device relative to a cab of the machine is determined using image information. A pose of the augmented reality device in a real world coordinate frame is determined using a pose of the machine in the real world coordinate frame and the pose of the augmented reality device relative to the cab of the machine. Digital content is provided on one or more displays of the augmented reality device. The digital content is arranged on the one or more displays based on the pose of the augmented reality device in the real world coordinate frame.
-
-
-
-
-
-
-
-
-