Abstract:
Methods and systems for interactive image analysis include receiving a selection of a region of an image and a request for analysis of the selection at an interface layer, transferring the selection and the request to an interpretation layer for analysis, dividing the selected region of the image into a plurality of sub-sections optimized for parallel computation to provide an analysis result that minimizes perceptible delay between receiving the request and receipt of results, analyzing the sub-sections using one or more execution nodes using a copy of the image stored in a shared memory, and providing combined analysis results to the interface layer for display.
Abstract:
A method for training a classifier to be operative as an epithelial texture classifier, includes obtaining a plurality of training micrograph areas of biopsy tissue and for each of the training micrograph areas, identifying probable locations of nuclei that form epithelia, generating a skeleton graph from the probable locations of the nuclei that form the epithelia, manually drawing walls on the skeleton graph outside of the epithelia to divide the epithelia from one another, and manually selecting points that lie entirely inside the epithelia to generate open and/or closed geodesic paths in the skeleton graph between pairs of the selected points. Data is obtained from points selected from the walls and the paths and applied to a classifier to train the classifier as the epithelial texture classifier. A method and detector for detecting epithelial structures includes applying a sample micrograph area of biopsy tissue to an epithelial texture classifier; identifying probable locations of nuclei that form epithelia of the sample micrograph area with the epithelial texture classifier, generating a skeleton graph from the probable locations of the nuclei that form the epithelia of the sample micrograph area, determining a set of open and/or closed geodesic paths in the skeleton graph of the sample micrograph area; and determining a set of the epithelial masks using the open and/or closed epithelial paths of the sample micrograph area.
Abstract:
A method and systems for cloud-based digital pathology include scanning received slides that include a pathology sample to produce a sample image in a shared memory, analyzing the sample image using one or more execution nodes, each including one or more processors, according to one or more analysis types to produce intermediate results, transmitting some or all of the sample image to a client device, further analyzing the sample image responsive to a request from the client device to produce a final analysis based on the intermediate results, and transmitting the final analysis to the client device.
Abstract:
Methods and apparatus for rendering a talking head on a client device are disclosed. The client device has a client cache capable of storing audio/visual data associated with rendering the talking head. The method comprises storing sentences in a client cache of a client device that relate to bridging delays in a dialog, storing sentence templates to be used in dialogs, generating a talking head response to a user inquiry from the client device, and determining whether sentences or stored templates stored in the client cache relate to the talking head response. If the stored sentences or stored templates relate to the talking head response, the method comprises instructing the client device to use the appropriate stored sentence or template from the client cache to render at least a part of the talking head response and transmitting a portion of the talking head response not stored in the client cache, if any, to the client device to render a complete talking head response. If the client cache has no stored data associated with the talking head response, the method comprises transmitting the talking head response to be rendered on the client device.
Abstract:
A method for training a classifier to be operative as an epithelial texture classifier, includes obtaining a plurality of training micrograph areas of biopsy tissue and for each of the training micrograph areas, identifying probable locations of nuclei that form epithelia, generating a skeleton graph from the probable locations of the nuclei that form the epithelia, manually drawing walls on the skeleton graph outside of the epithelia to divide the epithelia from one another, and manually selecting points that lie entirely inside the epithelia to generate open and/or closed geodesic paths in the skeleton graph between pairs of the selected points. Data is obtained from points selected from the walls and the paths and applied to a classifier to train the classifier as the epithelial texture classifier. A method and detector for detecting epithelial structures includes applying a sample micrograph area of biopsy tissue to an epithelial texture classifier; identifying probable locations of nuclei that form epithelia of the sample micrograph area with the epithelial texture classifier, generating a skeleton graph from the probable locations of the nuclei that form the epithelia of the sample micrograph area, determining a set of open and/or closed geodesic paths in the skeleton graph of the sample micrograph area; and determining a set of the epithelial masks using the open and/or closed epithelial paths of the sample micrograph area.
Abstract:
Disclosed is a parallel support vector machine technique for solving problems with a large set of training data where the kernel computation, as well as the kernel cache and the training data, are spread over a number of distributed machines or processors. A plurality of processing nodes are used to train a support vector machine based on a set of training data. Each of the processing nodes selects a local working set of training data based on data local to the processing node, for example a local subset of gradients. Each node transmits selected data related to the working set (e.g., gradients having a maximum value) and receives an identification of a global working set of training data. The processing node optimizes the global working set of training data and updates a portion of the gradients of the global working set of training data. The updating of a portion of the gradients may include generating a portion of a kernel matrix. These steps are repeated until a convergence condition is met. Each of the local processing nodes may store all, or only a portion of, the training data. While the steps of optimizing the global working set of training data, and updating a portion of the gradients of the global working set, are performed in each of the local processing nodes, the function of generating a global working set of training data is performed in a centralized fashion based on the selected data (e.g., gradients of the local working set) received from the individual processing nodes.
Abstract:
A method is provided for customizing a multi-media message created by a sender for a recipient, in which the multi-media message includes an animated entity audibly presenting speech converted from text by the sender. At least one image is received from the sender. Each of the at least one image is associated with a tag. The sender is presented with options to insert the tag associated with one of the at least one image into the sender text.
Abstract:
A system and method of controlling the movement of a virtual agent while the agent is listening to a human user during a conversation is disclosed. The method comprises receiving speech data from the user, performing a prosodic analysis of the speech data and controlling the virtual agent movement according to the prosodic analysis.
Abstract:
A system and method for generating photo-realistic talking-head animation from a text input utilizes an audio-visual unit selection process. The lip-synchronization is obtained by optimally selecting and concatenating variable-length video units of the mouth area. The unit selection process utilizes the acoustic data to determine the target costs for the candidate images and utilizes the visual data to determine the concatenation costs. The image database is prepared in a hierarchical fashion, including high-level features (such as a full 3D modeling of the head, geometric size and position of elements) and pixel-based, low-level features (such as a PCA-based metric for labeling the various feature bitmaps).
Abstract:
A method for generating animated sequences of talking heads in text-to-speech applications wherein a processor samples a plurality of frames comprising image samples. Representative parameters are extracted from the image samples and stored in an animation library. The processor also samples a plurality of multiphones comprising images together with their associated sounds. The processor extracts parameters from these images comprising data characterizing mouth shapes, maps, rules, or equations, and stores the resulting parameters and sound information in a coarticulation library. The animated sequence begins with the processor considering an input phoneme sequence, recalling from the coarticulation library parameters associated with that sequence, and selecting appropriate image samples from the animation library based on that sequence. The image samples are concatenated together, and the corresponding sound is output, to form the animated synthesis.