Abstract:
The invention is based, in part, on a system and method designed to be able to easily and automatically scale up to millions of cameras and users. To do this, this discourse teaches use of modern cloud computing technology, including automated service provisioning, automated virtual machine migration services, RESTful API, and various firewall traversing methods to facilitate the scaling process. Moreover, the system and method described herein teaches scalable cloud solutions providing for higher though-put camera provisioning and event recognition. The network may segregate the retrieval server from the storage server, and by doing so, minimizing the load on any one server and improving network efficiency and scalability
Abstract:
The invention is based, in part, on a system for allowing at least one client to real-time monitor and, or playback at least one real-world recognized event via at least one processor-controlled video camera, said system comprising: a processor; a non-transitory storage medium coupled to the processor; encoded instructions stored in the non-transitory storage medium, which when executed by the processor, cause the processor to: detect a threshold-grade event from audio-video data of a real-world environment captured from a processor-controlled video camera by an event detection module within an event management system applying event detection parameters; analyze the threshold-grade event for categorization into any one of a recognized event by an event recognition module within the event management system applying event recognition parameters; transmit at least any one of a single stream of the recognized event and, or a single stream of a audio-video sequence succeeding and, or preceding the recognized event to a client device; and facilitate at least any one of a user defined playback of the single stream of the recognized event, user defined monitoring of the audio-video sequence preceding and, or succeeding the recognized event, and, or remote provisioning of the processor-controlled video camera, whereby the playback and, or provisioning is facilitated via a client device user interface.
Abstract:
Artificial intelligence-based processing can be used to classify audio information received from an audio input unit. In an example, audio information can be received from a microphone configured to monitor an environment. A processor circuit can identify identifying one or more features of the audio information received from the microphone and use a first applied machine learning algorithm to analyze the one or more features and determine whether the audio information includes an indication of an abnormal event in the environment. In an example, the processor circuit can use a different second applied machine learning algorithm, such as a neural network-based deep learning algorithm, to analyze the same one or more features and classify the audio information as including an indication of a particular event type in the environment.
Abstract:
The invention is based, in part, on a system and method designed to be able to easily and automatically scale up to millions of cameras and users. To do this, this discourse teaches use of modern cloud computing technology, including automated service provisioning, automated virtual machine migration services, RESTful API, and various firewall traversing methods to facilitate the scaling process. Moreover, the system and method described herein teaches scalable cloud solutions providing for higher though-put camera provisioning and event recognition. The network may segregate the retrieval server from the storage server, and by doing so, minimizing the load on any one server and improving network efficiency and scalability.
Abstract:
The present invention discloses methods and systems for recognizing an object in an input image based on stored training images. An object recognition system the input image, computes a signature of the input image, compares the signature with one or more stored signatures and retrieves one or more matching images from the set of training images. The matching images are then displayed to the user for further action.
Abstract:
Method of tracking moveable objects (typically tagged objects that are moved by actors e.g. people, vehicles) by combining and analyzing data obtained from multiple types of sensors, such as video cameras, RFID tag readers, GPS sensors, and WiFi transceivers. Objects may be tagged by RFID tags, NFC tags, bar codes, or even tagged by visual appearance. The system operates in near real-time, and compensates for errors in sensor readings and missing sensor data by modeling object and actor movement according to a plurality of possible paths, weighting data from some sensors higher than others according to estimates of sensor accuracy, and weighing the probability of certain paths according to various other rules and penalty cost parameters. The system can maintain a comprehensive database which can be queried as to which actors associate with which objects, and vice versa. Other data pertaining to object location and association can also be obtained.
Abstract:
A system and method for detecting human presence in or absence from a field-of-view of a camera by analyzing camera data using a processor inside of or adjacent to the camera itself. In an example, the camera can be integrated with or embedded in another edge-based sensor device. In an example, a video signal processing system receives image data from one or more image sensors and uses a local processing circuit to process the image data and determine if a human being is or is not present during a particular time, interval, or sequence of frames. In an example, the human being identification technique can be used in security or surveillance applications such as for home, business, or other monitoring cameras.
Abstract:
Artificial intelligence-based processing can be used to classify audio information received from an audio input unit. In an example, audio information can be received from a microphone configured to monitor an environment. A processor circuit can identify identifying one or more features of the audio information received from the microphone and use a first applied machine learning algorithm to analyze the one or more features and determine whether the audio information includes an indication of an abnormal event in the environment. In an example, the processor circuit can use a different second applied machine learning algorithm, such as a neural network-based deep learning algorithm, to analyze the same one or more features and classify the audio information as including an indication of a particular event type in the environment.
Abstract:
The invention is based, in part, on a system for allowing at least one client to real-time monitor and, or playback at least one real-world recognized event via at least one processor-controlled video camera, said system comprising: a processor; a non-transitory storage medium coupled to the processor; encoded instructions stored in the non-transitory storage medium, which when executed by the processor, cause the processor to: detect a threshold-grade event from audio-video data of a real-world environment captured from a processor-controlled video camera by an event detection module within an event management system applying event detection parameters; analyze the threshold-grade event for categorization into any one of a recognized event by an event recognition module within the event management system applying event recognition parameters; transmit at least any one of a single stream of the recognized event and, or a single stream of a audio-video sequence succeeding and, or preceding the recognized event to a client device; and facilitate at least any one of a user defined playback of the single stream of the recognized event, user defined monitoring of the audio-video sequence preceding and, or succeeding the recognized event, and, or remote provisioning of the processor-controlled video camera, whereby the playback and, or provisioning is facilitated via a client device user interface.
Abstract:
Method of tracking moveable objects (typically tagged objects that are moved by actors e.g. people, vehicles) by combining and analyzing data obtained from multiple types of sensors, such as video cameras, RFID tag readers, GPS sensors, and WiFi transceivers. Objects may be tagged by RFID tags, NFC tags, bar codes, or even tagged by visual appearance. The system operates in near real-time, and compensates for errors in sensor readings and missing sensor data by modeling object and actor movement according to a plurality of possible paths, weighting data from some sensors higher than others according to estimates of sensor accuracy, and weighing the probability of certain paths according to various other rules and penalty cost parameters. The system can maintain a comprehensive database which can be queried as to which actors associate with which objects, and vice versa. Other data pertaining to object location and association can also be obtained.