Abstract:
A system and method for object tracking and timing across multiple camera views includes local and global tracking modules for tracking the location of objects as they traverse particular regions of interest within an area of interest. A local timing module measures the time spent with each object within the area captured by a camera. A global timing module measures the time taken by the tracked object to traverse the entire area of interest or the length of the stay of the object within the area of interest.
Abstract:
This disclosure provides a method and system for automated sequencing of vehicles in side-by-side drive-thru configurations via appearance-based classification. According to an exemplary embodiment, an automated sequencing method includes computer-implemented method of automated sequencing of vehicles in a side-by-side drive-thru, the method comprising: a) an image capturing device capturing video of a merge-point area associated with multiple lanes of traffic merging; b) detecting in the video a vehicle as it traverses the merge-point area; c) classifying the detected vehicle associated with traversing the merge-point area as coming from one of the merging lanes; and d) aggregating vehicle classifications performed in step c) to generate a merge sequence of detected vehicles.
Abstract:
A method for updating an event sequence includes acquiring video data of a queue area from at least one image source; searching the frames for subjects located at least near a region of interest (ROI) of defined start points in the video data; tracking a movement of each detected subject through the queue area over a subsequent series of frames; using the tracking, determining if a location of the a tracked subject reaches a predefined merge point where multiple queues in the queue area converge into a single queue lane; in response to the tracked subject reaching the predefined merge point, computing an observed sequence of where the tracked subject places among other subjects approaching an end-event point; and, updating a sequence of end-events to match the observed sequence of subjects in the single queue lane.
Abstract:
A method for computing output using a non-contact (invisible) input signal includes acquiring depth data of a scene captured by a depth-capable sensor. The method includes generating a temporal series of depth maps corresponding to the depth data. The method includes generating at least one volumetric attribute from the depth data. The method includes generating an output based on the volumetric attribute to control actions.
Abstract:
A computer-vision based method for validating an activity workflow of a human performer includes identifying a target activity. The method includes determining an expected sequence of actions associated with the target activity. The method includes receiving a video stream from an image capture device monitoring an activity performed by an associated human performer. The method includes determining an external cue in the video stream. The method includes associating a frame capturing the external cue as a first frame in a key frame sequence. The method includes determining an action being performed by the associated human performer in the key frame sequence. In response to determining the action in the key frame sequence matching an expected action in the target activity, the method includes verifying the action as being performed in the monitored activity. In response to not determining the action in the key frame sequence, the method includes generating an alert indicating an error in the monitored activity.
Abstract:
A method for reconstructing an image of a scene captured using a compressed sensing device. A mask is received which identifies at least one region of interest in an image of a scene. Measurements are then obtained of the scene using a compressed sensing device comprising, at least in part, a spatial light modulator configuring a plurality of spatial patterns according to a set of basis functions each having a different spatial resolution. A spatial resolution is adaptively modified according to the mask. Each pattern focuses incoming light of the scene onto a detector which samples sequential measurements of light. These measurements comprise a sequence of projection coefficients corresponding to the scene. Thereafter, an appearance of the scene is reconstructed utilizing a compressed sensing framework which reconstructs the image from the sequence of projection coefficients.
Abstract:
A system and method for optimizing video-based tracking of an object of interest are provided. A video of a regularized motion environment that comprise multiple video frames is acquired and an initial instance of an object of interest in one of the frames is then detected. An expected size and orientation of the object of interest as a function of the location of the object is then determined. The location of the object of interest is then determined in a next subsequent frame using the expected size and orientation of the object of interest.
Abstract:
A camera outputs video as a sequence of video frames having pixel values in a first (e.g., relatively low dimensional) color space, where the first color space has a first number of channels. An image-processing device maps the video frames to a second (e.g., relatively higher dimensional) color representation of video frames. The mapping causes the second color representation of video frames to have a greater number of channels relative to the first number of channels. The image-processing device extracts a second color representation of a background frame of the scene. The image-processing device can then detect foreground objects in a current frame of the second color representation of video frames by comparing the current frame with the second color representation of a background frame. The image-processing device then outputs an identification of the foreground objects in the current frame of the video.
Abstract:
What is disclosed is system and method for contemporaneously reconstructing images of a scene illuminated with unstructured and structured illumination sources. In one embodiment, the system comprises capturing a first 2D image containing energy reflected from a scene being illuminated by a structured illumination source and a second 2D image containing energy reflected from the scene being illuminated by an unstructured illumination source. A controller effectuates a manipulation of the structured and unstructured illumination sources during capture of the video. A processor is configured to execute machine readable program instructions enabling the controller to manipulate the illumination sources, and for effectuating the contemporaneous reconstruction of a 2D intensity map of the scene using the second 2D image and of a 3D surface map of the scene using the first 2D image. The reconstruction is effectuated by manipulating the illumination sources.
Abstract:
What is disclosed is a wireless cellular device capable of determining a volume of an object in an image captured by a camera of that apparatus. In one embodiment, the present wireless cellular device comprises an illuminator for projecting a pattern of structured light with known spatial characteristics, and a camera for capturing images of an object for which a volume is to be estimated. The camera is sensitive to a wavelength range of the projected pattern of structured light. A spatial distortion is introduced by a reflection of the projected pattern off a surface of the object. And processor executing machine readable program instructions for performing the method of: receiving an image of the object from the camera; processing the image to generate a depth map; and estimating a volume of the object from the depth map. A method for using the present wireless cellular device is also provided.