SYSTEM AND METHOD FOR PHYSICAL MANIPULATION OF OBJECT AND TYPE OF OBJECT AS INPUT
    31.
    发明申请
    SYSTEM AND METHOD FOR PHYSICAL MANIPULATION OF OBJECT AND TYPE OF OBJECT AS INPUT 审中-公开
    用于对象和输入对象类型的物理操作的系统和方法

    公开(公告)号:US20150253859A1

    公开(公告)日:2015-09-10

    申请号:US14197798

    申请日:2014-03-05

    Applicant: Google Inc.

    CPC classification number: G06F3/017 G06F3/0304

    Abstract: A system and method is provided of detecting user manipulation of an inanimate object and interpreting that manipulation as input. In one aspect, the manipulation may be detected by an image capturing component of a computing device, and the manipulation is interpreted as an instruction to execute a command, such as opening up a drawing application in response to a user picking up a pen. The manipulation may also be detected with the aid of an audio capturing device, e.g., a microphone on the computing device.

    Abstract translation: 提供了一种系统和方法,用于检测无用物体的用户操纵,并将该操作解释为输入。 在一个方面,可以通过计算设备的图像捕获部件检测操作,并且操作被解释为执行命令的指令,诸如响应于用户拿起笔而打开绘图应用程序。 借助于诸如计算设备上的麦克风的音频捕获设备,也可以检测该操作。

    3D model updates using crowdsourced video
    32.
    发明授权
    3D model updates using crowdsourced video 有权
    使用众包视频的3D模型更新

    公开(公告)号:US08818081B1

    公开(公告)日:2014-08-26

    申请号:US14055132

    申请日:2013-10-16

    Applicant: Google Inc.

    Abstract: An exemplary method includes prompting a user to capture video data at a location. The location is associated with navigation directions for the user. Information representing visual orientation and positioning information associated with the captured video data is received by one or more computing devices, and a stored data model representing a 3D geometry depicting objects associated with the location is accessed. Between corresponding images from the captured video data and projections of the 3D geometry, one or more candidate change regions are detected. Each candidate change region indicates an area of visual difference between the captured video data and projections. When it is detected that a count of the one or more candidate change regions is below a threshold, the stored model data is updated with at least part of the captured video data based on the visual orientation and positioning information associated with the captured video data.

    Abstract translation: 一种示例性方法包括提示用户捕获位置处的视频数据。 该位置与用户的导航方向相关联。 表示与捕获的视频数据相关联的视觉定向和定位信息的信息由一个或多个计算设备接收,并且存储表示描绘与该位置相关联的对象的3D几何的存储的数据模型。 在捕获的视频数据和3D几何的投影的相应图像之间,检测一个或多个候选变化区域。 每个候选变化区域表示所捕获的视频数据和投影之间的视差的区域。 当检测到一个或多个候选变化区域的计数低于阈值时,基于与捕获的视频数据相关联的视觉取向和定位信息,利用所捕获的视频数据的至少一部分来更新存储的模型数据。

    Depth map generation
    33.
    发明授权
    Depth map generation 有权
    深度图生成

    公开(公告)号:US08760500B1

    公开(公告)日:2014-06-24

    申请号:US14061423

    申请日:2013-10-23

    Applicant: Google Inc.

    Abstract: Aspects of the disclosure relate generally to generating depth data from a video. As an example, one or more computing devices may receive an initialization request for a still image capture mode. After receiving the request to initialize the still image capture mode, the one or more computing devices may automatically begin to capture a video including a plurality of image frames. The one or more computing devices track features between a first image frame of the video and each of the other image frames of the video. Points corresponding to the tracked features may be generated by the one or more computing devices using a set of assumptions. The assumptions may include a first assumption that there is no rotation and a second assumption that there is no translation. The one or more computing devices then generate a depth map based at least in part on the points.

    Abstract translation: 本公开的方面通常涉及从视频生成深度数据。 作为示例,一个或多个计算设备可以接收静态图像捕获模式的初始化请求。 在接收到初始化静止图像捕获模式的请求之后,一个或多个计算设备可以自动开始捕获包括多个图像帧的视频。 一个或多个计算设备跟踪视频的第一图像帧和视频的每个其他图像帧之间的特征。 与被跟踪的特征相对应的点可以由一个或多个计算设备使用一组假设来生成。 这些假设可以包括没有旋转的第一假设和没有平移的第二假设。 然后,一个或多个计算设备至少部分地基于点生成深度图。

    Time-lapsed image sequence generation

    公开(公告)号:US09607243B1

    公开(公告)日:2017-03-28

    申请号:US14591256

    申请日:2015-01-07

    Applicant: Google Inc.

    CPC classification number: G06T5/10 G06K9/00677 G06K9/6255 G06T13/80

    Abstract: Aspects of the disclosure relate to providing users with sequences of images of physical locations over time or time-lapses. In order to do so, a set of images of a physical location may be identified. From the set of images, a representative image may be selected. The set may then be filtered by comparing the other images in the set to the representative image. The images in the filtered set may then be aligned to the representative image. From this set, a time-lapsed sequence of images may be generated, and the amount of change in the time-lapsed sequence of images may be determined. At the request of a user device for a time-lapsed image representation of the specified physical location, the generated time-lapsed sequence of images may be provided.

    CAPTURE AND RENDER OF PANORAMIC VIRTUAL REALITY CONTENT
    38.
    发明申请
    CAPTURE AND RENDER OF PANORAMIC VIRTUAL REALITY CONTENT 审中-公开
    全景虚拟现实内容的捕捉和渲染

    公开(公告)号:US20160353089A1

    公开(公告)日:2016-12-01

    申请号:US14723151

    申请日:2015-05-27

    Applicant: GOOGLE INC.

    Abstract: Systems and methods are described for defining a set of images based on captured images, receiving a viewing direction associated with a user of a virtual reality (VR) head mounted display, receiving an indication of a change in the viewing direction. The methods further include configuring, a re-projection of a portion of the set of images, the re-projection based at least in part on the changed viewing direction and a field of view associated with the captured images, and converting the portion from a spherical perspective projection into a planar perspective projection, rendering by the computing device and for display in the VR head mounted display, an updated view based on the re-projection, the updated view configured to correct distortion and provide stereo parallax in the portion, and providing, to the head mounted display, the updated view including a stereo panoramic scene corresponding to the changed viewing direction.

    Abstract translation: 描述了基于捕获的图像来定义一组图像的系统和方法,接收与虚拟现实(VR)头戴式显示器的用户相关联的观看方向,接收观看方向的改变的指示。 所述方法还包括至少部分地基于所改变的观察方向和与所捕获图像相关联的视场来配置,所述图像集合的一部分的重新投影,所述重新投影将所述部分从 球面透视投影到平面透视投影中,由计算设备渲染并用于显示在VR头戴式显示器中,基于重新投影的更新视图,配置为校正失真并在该部分中提供立体视差的更新视图,以及 向头戴式显示器提供包括与改变的观看方向对应的立体全景场景的更新视图。

    AUTOMATIC DETECTION OF PANORAMIC GESTURES
    39.
    发明申请
    AUTOMATIC DETECTION OF PANORAMIC GESTURES 有权
    自动检测全景镜

    公开(公告)号:US20160344930A1

    公开(公告)日:2016-11-24

    申请号:US14717492

    申请日:2015-05-20

    Applicant: Google Inc.

    Abstract: Aspects of the disclosure relate to capturing panoramic images using a computing device. For example, the computing device may record a set of video frames and tracking features each including one or more features that appear in two or more video frames of the set of video frames within the set of video frames may be determined. A set of frame-based features based on the displacement of the tracking features between two or more video frames of the set of video frames may be determined by the computing device. A set of historical feature values based on the set of frame-based features may also be determined by the computing device. The computing device may determine then whether a user is attempting to capture a panoramic image based on the set of historical feature values. In response, the computing device may capture a panoramic image.

    Abstract translation: 本公开的方面涉及使用计算设备捕获全景图像。 例如,计算设备可以记录一组视频帧,并且可以确定在视频帧集合内出现在视频帧集合中的两个或更多个视频帧中的包括一个或多个特征的跟踪特征。 基于视频帧集合中的两个或多个视频帧之间的跟踪特征的位移的基于帧的特征的集合可以由计算设备确定。 基于基于帧的特征集合的一组历史特征值也可以由计算设备确定。 计算设备可以确定用户是否正在尝试基于该组历史特征值捕获全景图像。 作为响应,计算设备可以捕获全景图像。

    THREE DIMENSIONAL NAVIGATION AMONG PHOTOS

    公开(公告)号:US20160307370A1

    公开(公告)日:2016-10-20

    申请号:US15195155

    申请日:2016-06-28

    Applicant: Google Inc.

    Abstract: An exemplary method for navigating among photos includes determining, using one or more computing devices, visual characteristics of a person depicted in a first image associated with a first location. These visual characteristics of the person are detected in a second image associated with a second location. Using the one or more computing devices, a series of intermediate images are identified based on the first location and the second location. Each intermediate image is associated with a location. The series of intermediate images and the second image are provided. Images of an intermediate destination from the series of intermediate images are selected based on a density of images at the intermediate destination. A 3D reconstruction of the intermediate destination is then generated based on the selected images. Thereafter, a visual presentation of images traversing through the 3D reconstruction of the intermediate destination to the second image is prepared for display.

Patent Agency Ranking