摘要:
A gesture recognition interface for use in controlling self-service machines and other devices is disclosed. A gesture is defined as motions and kinematic poses generated by humans, animals, or machines. Specific body features are tracked, and static and motion gestures are interpreted. Motion gestures are defined as a family of parametrically delimited oscillatory motions, modeled as a linear-in-parameters dynamic system with added geometric constraints to allow for real-time recognition using a small amount of memory and processing time. A linear least squares method is preferably used to determine the parameters which represent each gesture. Feature position measure is used in conjunction with a bank of predictor bins seeded with the gesture parameters, and the system determines which bin best fits the observed motion. Recognizing static pose gestures is preferably performed by localizing the body/object from the rest of the image, describing that object, and identifying that description. The disclosure details methods for gesture recognition, as well as the overall architecture for using gesture recognition to control of devices, including self-service machines.
摘要:
An intelligent system for automatically monitoring, diagnosing, and repairing complex hardware and software systems is presented. A number of functional modules enable the system to collect relevant data from both hardware and software components, analyze the incoming data to detect faults, further monitor sensor data and historical knowledge to predict potential faults, determine an appropriate response to fix the faults, and finally automatically repair the faults when appropriate. The system leverages both software and hardware modules to interact with the complex system being monitored. Additionally, the lessons learned on one system can be applied to better understand events occurring on the same or similar systems.
摘要:
A method for controlling a graphical user interface (GUI) for a touchscreen-enabled computer systems provides a variety of software methods (tools) provide for high-fidelity control of the user interface. The TrackScreen tool provides finger-friendly mouse functions such as scrolling, dragging and clicking. The Magnifier application continuously captures the current screen image, and displays a magnified subset of it. Selecting within this magnified area with a pointing device (mouse, touchscreen, digitizer, etc) causes the application to simulate the action on the portion of the screen corresponding to the point in the magnified image that was selected. A KeyBoard application, a keyboard is rendered on screen, with sufficient size that the individual keys are easily selectable with an unaided finger. The Common Tasks Tool or CTT) allows common keyboard shortcuts, mouse events, and other user interface events to be specified in a configuration file and represented on screen as a large, easy-to-click button. The Touchscreen Task Switcher is invoked using any interface (software or hardware) element, and visually takes up the entire screen. The Touchscreen Snapshot utility ties in with an external camera with a physical button on it. The Window Template Manager (WTM), is used to specify, and then instantiate, the position and sizes of multiple windows for use with a touchscreen display. The Touch Portal is a full-screen application with a set of customizable buttons representing applications and other tools.