Abstract:
Methods, systems, computer program products, and methods of doing business by providing bookmarks for voice mail messages in order to facilitate improved navigation thereof. A bookmark is preferably conveyed by the caller leaving a voice mail message to indicate information such as a change in the topic of the message, or a particular important passage within the message, or perhaps to identify some information for special treatment (such as a telephone number and time of day at which the call can be returned). Once the message has been bookmarked, the listener can navigate the message more efficiently (for example, by listening to a few seconds of each topic and then skipping to the next topic, or by listening only to the special information) and can also perform actions on parts of the voice mail message (such as forwarding only a selected segment to a third party, while keeping the remainder of the message confidential). Optionally, audio cues may be rendered with the voice mail message as it is played to the listener, where audio cues are sounds that are either “incorporated in” with the audio rendering as a separate (background) audio stream or inserted in-line within the message. (For example, audio cues may change to signify that the bookmarked topics of the message change, or many announce the type of bookmark associated with an upcoming audio message segment).
Abstract:
A video-conferencing device for presenting augmented images that includes at least one interface, a network and a computer processor programmed to receive first information identifying a scene via the at least one interface. The computer processor also detects whether the scene contains at least one marker and identifies a location of each detected marker within the scene. In response to determining that the scene contains a first marker and based on the location of the first marker, the computer processor then augments the portion of the scene containing the first marker with second information. The computer processor then transmits the augmented scene to at least one external device via the network.
Abstract:
A method and system for processing messages within the framework of an integrated message system. Recipients of messages in an integrated messaging system are provided with an authentic impression of the received message. In a first step, a message received within the framework of an integrated messaging system is automatically translated. Language detection and dictation system is provided. The message contents of the incoming message as well as its segments and parameters are simultaneously utilized to generate additional information regarding the sender and the information, which is suitable to give the recipient an impression of the received message in the most authentic form possible.
Abstract:
Aspects of the present invention are directed at obtaining contextual information with a voicemail message. In accordance with one embodiment, a method is provided that obtains additional contextual information that is not obtained automatically when a voicemail message is received. More specifically, the method includes automatically obtaining a first set of contextual information from a client associated with the caller when the caller is transferred to a voicemail system. Then a determination is made regarding whether the callee requests that additional contextual information be obtained. If the callee requests that additional contextual information be obtained, the requested information is obtained from a third-party service or by prompting the caller.
Abstract:
In one embodiment, techniques for manipulating a voice message using a telephone user interface are provided. The telephone user interface receives a mark begin segment command from a telephonic device through a voice modality. The mark begin segment command indicates a beginning point for a voice message segment in the voice message. The telephone user interface then receives a mark end segment command from the telephonic device through the voice modality. The mark segment command indicates an ending point for the voice message segment in the voice message. An action command from the telephonic device is received and an action based on the action command for the voice message segment is performed. The voice message segment is a portion of the voice message determined based on the indicated beginning point of the voice message segment and the indicated ending point of the voice message segment in the voice message.
Abstract:
Asynchronous media communications consist of sending text or voice messages with tags indicating the priority of the message. AMC combines voice messaging with features such as tagging and threading of voice strings on a recipient device. Communication can take place on a one-to-one or one-to-many basis.
Abstract:
This apparatus calculates, from the spoken dialogue data between a customer and an operator, the speech length of each speech. Then, it estimates the structure of the spoken dialogue data based on the speech length and specifies a speech-length-characteristic change time point (time) when the speaker who makes longer speeches changes. Then it extracts an information search time point (time) when the operator performs information search from operation log information about operator's operations performed during this conversation and determined both the ending time point of a part-to-elicit from the information search time point and the starting time point of the part-to-elicit from the speech length-characteristic change time point.
Abstract:
A system for producing a transcript of a meeting having n attendees, the attendees being identified as ID1 to IDn and channel 1 to channel n. A speech discriminator includes a channel monitor which generates a speech output from one or more of channels 1 to n at any one times a speech file selector at 14 and a speech file database at 15. Discrimination is on the basis of pre-allocated channels which correspond to pre-allocated microphones which are matched by ID and to the speech files in the speech file database. The effect of 13, 14 and 15 is to match a channel input to a particular speech file in the database 15 so that this information may then be passed to the audio to text convertor such that the speech file information and the input audio may be converted to text, displayed and written to a text file.