COMBINING MOBILE DEVICES WITH PEOPLE TRACKING FOR LARGE DISPLAY INTERACTIONS

    公开(公告)号:US20170090560A1

    公开(公告)日:2017-03-30

    申请号:US14866534

    申请日:2015-09-25

    Abstract: The large display interaction implementations described herein combine mobile devices with people tracking to enable new interactions including making a non-touch-sensitive display touch-sensitive and allowing personalized interactions with the display. One implementation tracks one or more mobile computing device users relative to a large computer-driven display, and configures content displayed on the display based on a distance a given mobile computing device user is from the display. Another implementation personalizes user interactions with a large display. One or more mobile computing device users are tracked relative to a display. The identity of each of the one or more mobile computing device users is obtained. Content displayed on the display is configured based on a distance an identified mobile computing device user is from the display and the identity of the user that provides the content.

    GAZE TRACKING VIA EYE GAZE MODEL
    22.
    发明申请
    GAZE TRACKING VIA EYE GAZE MODEL 有权
    通过眼睛大小模型的GAZE跟踪

    公开(公告)号:US20160202756A1

    公开(公告)日:2016-07-14

    申请号:US14593955

    申请日:2015-01-09

    CPC classification number: G06F3/013 G06F3/0304 G06F3/038 G06K9/0061

    Abstract: Examples are disclosed herein that are related to gaze tracking via image data. One example provides, on a gaze tracking system comprising an image sensor, a method of determining a gaze direction, the method comprising acquiring image data via the image sensor, detecting in the image data facial features of a human subject, determining an eye rotation center based upon the facial features using a calibrated face model, determining an estimated position of a center of a lens of an eye from the image data, determining an optical axis based upon the eye rotation center and the estimated position of the center of the lens, determining a visual axis by applying an adjustment to the optical axis, determining the gaze direction based upon the visual axis, and providing an output based upon the gaze direction.

    Abstract translation: 本文公开了与经由图像数据的凝视跟踪有关的示例。 一个示例在包括图像传感器的注视跟踪系统,确定注视方向的方法中提供,该方法包括经由图像传感器获取图像数据,在图像数据中检测人类对象的面部特征,确定眼睛旋转中心 基于使用校准面部模型的面部特征,根据图像数据确定眼睛的眼睛的中心的估计位置,基于眼睛旋转中心和眼镜的中心的估计位置来确定光轴, 通过对光轴施加调整来确定视轴,基于视轴确定注视方向,以及基于注视方向提供输出。

    Dynamic matrix convolution with channel fusion

    公开(公告)号:US12223412B2

    公开(公告)日:2025-02-11

    申请号:US17123697

    申请日:2020-12-16

    Abstract: A computer device for automatic feature detection comprises a processor, a communication device, and a memory configured to hold instructions executable by the processor to instantiate a dynamic convolution neural network, receive input data via the communication network, and execute the dynamic convolution neural network to automatically detect features in the input data. The dynamic convolution neural network compresses the input data from an input space having a dimensionality equal to a predetermined number of channels into an intermediate space having a dimensionality less than the number of channels. The dynamic convolution neural network dynamically fuses the channels into an intermediate representation within the intermediate space and expands the intermediate representation from the intermediate space to an expanded representation in an output space having a higher dimensionality than the dimensionality of the intermediate space. The features in the input data are automatically detected based on the expanded representation.

    Calibrating cameras using human skeleton

    公开(公告)号:US11847796B2

    公开(公告)日:2023-12-19

    申请号:US17249049

    申请日:2021-02-18

    CPC classification number: G06T7/80 G06T7/70

    Abstract: Examples are disclosed herein that relate to automatically calibrating cameras based on human detection. One example provides a computing system comprising instructions executable to receive image data comprising depth image data and two-dimensional image data of a space from a camera, detect a person in the space via the image data, determine a skeletal representation for the person via the image data, determine over a period of time a plurality of locations at which a reference point of the skeletal representation is on a ground area in the image data, determine a ground plane of the three-dimensional representation based upon the plurality of locations at which the reference point of the skeletal representation is on the ground area in the image data, and track a location of an object within the space relative to the ground plane.

Patent Agency Ranking