Abstract:
A method and system for detecting temporal segments of talking faces in a video sequence using visual cues. The system detects talking segments by classifying talking and non-talking segments in a sequence of image frames using visual cues. The present disclosure detects temporal segments of talking faces in video sequences by first localizing face, eyes, and hence, a mouth region. Then, the localized mouth regions across the video frames are encoded in terms of integrated gradient histogram (IGH) of visual features and quantified using evaluated entropy of the IGH. The time series data of entropy values from each frame is further clustered using online temporal segmentation (K-Means clustering) algorithm to distinguish talking mouth patterns from other mouth movements. Such segmented time series data is then used to enhance the emotion recognition system.
Abstract:
A method for converting an electronic flash storage device having a byte addressable storage (ByAS) and a block addressable flash storage (BlAS) to a single byte addressable storage includes receiving, by a host, a request for memory allocation from the ByAS, the receiving being from a first application among of a plurality of applications running on a processor; deallocating, by the host, a least relevant page allocated to at least one second application among the plurality of applications; moving, by the host, a content to the BlAS at a first BlAS location, the content related to the least relevant page, the moving based on the deallocation; allocating, by the host, the least relevant page to the first application; and updating, by the host, a cache metadata and a page lookup table of the first application and the at least one second application based on the deallocation and allocation.
Abstract:
A method and system for detecting temporal segments of talking faces in a video sequence using visual cues. The system detects talking segments by classifying talking and non-talking segments in a sequence of image frames using visual cues. The present disclosure detects temporal segments of talking faces in video sequences by first localizing face, eyes, and hence, a mouth region. Then, the localized mouth regions across the video frames are encoded in terms of integrated gradient histogram (IGH) of visual features and quantified using evaluated entropy of the IGH. The time series data of entropy values from each frame is further clustered using online temporal segmentation (K-Means clustering) algorithm to distinguish talking mouth patterns from other mouth movements. Such segmented time series data is then used to enhance the emotion recognition system.