Methods and systems for enabling human-robot interaction to resolve task ambiguity

    公开(公告)号:US11501777B2

    公开(公告)日:2022-11-15

    申请号:US17161767

    申请日:2021-01-29

    Abstract: The disclosure herein relates to methods and systems for enabling human-robot interaction (HRI) to resolve task ambiguity. Conventional techniques that initiates continuous dialogue with the human to ask a suitable question based on the observed scene until resolving the ambiguity are limited. The present disclosure use the concept of Talk-to-Resolve (TTR) which initiates a continuous dialogue with the user based on visual uncertainty analysis and by asking a suitable question that convey the veracity of the problem to the user and seek guidance until all the ambiguities are resolved. The suitable question is formulated based on the scene understanding and the argument spans present in the natural language instruction. The present disclosure asks questions in a natural way that not only ensures that the user can understand the type of confusion, the robot is facing; but also ensures minimal and relevant questioning to resolve the ambiguities.

    Method and system for prediction of correct discrete sensor data based on temporal uncertainty

    公开(公告)号:US11429467B2

    公开(公告)日:2022-08-30

    申请号:US16728528

    申请日:2019-12-27

    Abstract: This disclosure relates generally to a method and system for prediction of correct discrete sensor data, thus enabling continuous flow of data even when a discrete sensor fails. The activities of humans/subjects, housed in a smart environment is continuously monitored by plurality of non-intrusive discrete sensors embedded in living infrastructure. The collected discrete sensor data is usually sparse and largely unbalanced, wherein most of the discrete sensor data is ‘No’ and comparatively only a few samples of ‘Yes’, hence making prediction very challenging. The proposed prediction techniques based on introduction of temporal uncertainty is performed in several stages which includes pre-processing of received discrete sensor data, introduction of temporal uncertainty techniques followed by prediction based on neural network techniques of learning pattern using historical data.

    Method and system for generating face animations from speech signal input

    公开(公告)号:US11295501B1

    公开(公告)日:2022-04-05

    申请号:US17188512

    申请日:2021-03-01

    Abstract: Most of the prior art references that generate animations fail to determine and consider head movement data. The prior art references which consider the head movement data for generating the animations rely on a sample video to generate/determine the head movements data, which, as a result, fail to capture changing head motions throughout course of a speech given by a subject in an actual whole length video. The disclosure herein generally relates to generating facial animations, and, more particularly, to a method and system for generating the facial animations from speech signal of a subject. The system determines the head movement, lip movements, and eyeball movements, of the subject, by processing a speech signal collected as input, and uses the head movement, lip movements, and eyeball movements, to generate an animation.

    Identity preserving realistic talking face generation using audio speech of a user

    公开(公告)号:US11176724B1

    公开(公告)日:2021-11-16

    申请号:US17036583

    申请日:2020-09-29

    Abstract: Speech-driven facial animation is useful for a variety of applications such as telepresence, chatbots, etc. The necessary attributes of having a realistic face animation are: 1) audiovisual synchronization, (2) identity preservation of the target individual, (3) plausible mouth movements, and (4) presence of natural eye blinks. Existing methods mostly address audio-visual lip synchronization, and synthesis of natural facial gestures for overall video realism. However, existing approaches are not accurate. Present disclosure provides system and method that learn motion of facial landmarks as an intermediate step before generating texture. Person-independent facial landmarks are generated from audio for invariance to different voices, accents, etc. Eye blinks are imposed on facial landmarks and the person-independent landmarks are retargeted to person-specific landmarks to preserve identity related facial structure. Facial texture is then generated from person-specific facial landmarks that helps to preserve identity-related texture.

    System and method for analyzing gait and postural balance of a person

    公开(公告)号:US11033205B2

    公开(公告)日:2021-06-15

    申请号:US15427762

    申请日:2017-02-08

    Abstract: A method and system is provided for finding and analyzing gait parameters and postural balance of a person using a Kinect system. The system is easy to use and can be installed at home as well as in clinic. The system includes a Kinect sensor, a software development kit (SDK) and a processor. The temporal skeleton information obtained from the Kinect sensor to evaluate gait parameters includes stride length, stride time, stance time and swing time. Eigenvector based curvature detection is used to analyze the gait pattern with different speeds. In another embodiment, Eigenvector based curvature detection is employed to detect static single limb stance (SLS) duration along with gait variables for evaluating body balance.

    Face video based heart rate monitoring using pulse signal modelling and tracking

    公开(公告)号:US10755125B2

    公开(公告)日:2020-08-25

    申请号:US15900788

    申请日:2018-02-20

    Abstract: The present disclosure provides a non-invasive, inexpensive and unobtrusive system that enables heart rate (HR) monitoring by addressing the traditionally known issues with face video based systems due to respiration, facial expressions, out-of-plane movements, camera parameters and environmental factors. These issues are alleviated by filtering, pulse modelling and HR tracking. Quality measures are defined which incorporate out-of-plane movements to define the quality of each video frame unlike existing approaches which provide a single quality for the entire video. To handle out-of-plane movement, Fourier basis function is employed to reconstruct pulse signals at affected locations. Bayesian decision theory based method performs HR tracking using previous HR and quality estimates for improved HR monitoring.

    Heart rate estimation from face videos using quality based fusion

    公开(公告)号:US10750959B2

    公开(公告)日:2020-08-25

    申请号:US15872458

    申请日:2018-01-16

    Abstract: A system and method for real time estimation of heart rate (HR) from one or more face videos acquired in non-invasive manner. The system receives face videos and obtains several blocks as ROI consisting of facial skin areas. Subsequently, the temporal fragments are extracted from the blocks and filtered to minimize the noise. In the next stage, several temporal fragments are extracted from the video. The several temporal fragments, corrupted by noise are determined using an image processing range filter and pruned for further processing. The HR of each temporal fragment, referred as local HR is estimated along with its quality. Eventually, a quality based fusion is applied to estimate a global HR corresponding to the received face videos. In addition, the disclosure herein is also applicable for frontal, profile and multiple faces and performs in real-time.

Patent Agency Ranking