Providing expressive user interaction with a multimodal application
    1.
    发明授权
    Providing expressive user interaction with a multimodal application 有权
    提供与多模式应用程序的表达性用户交互

    公开(公告)号:US08725513B2

    公开(公告)日:2014-05-13

    申请号:US11734422

    申请日:2007-04-12

    CPC分类号: G10L15/22

    摘要: Methods, apparatus, and products are disclosed for providing expressive user interaction with a multimodal application, the multimodal application operating in a multimodal browser on a multimodal device supporting multiple modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine through a VoiceXML interpreter, including: receiving, by the multimodal browser, user input from a user through a particular mode of user interaction; determining, by the multimodal browser, user output for the user in dependence upon the user input; determining, by the multimodal browser, a style for the user output in dependence upon the user input, the style specifying expressive output characteristics for at least one other mode of user interaction; and rendering, by the multimodal browser, the user output in dependence upon the style.

    摘要翻译: 公开了方法,装置和产品,用于提供与多模式应用程序的表达性用户交互,多模式应用程序在支持多种模式的用户交互的多模式设备上的多模式浏览器中操作,包括语音模式和一种或多种非语音模式, 多模式应用通过VoiceXML解释器可操作地耦合到语音引擎,包括:由多模式浏览器通过用户交互的特定模式从用户接收用户输入; 由所述多模式浏览器根据所述用户输入确定所述用户的用户输出; 根据所述用户输入,通过所述多模式浏览器确定所述用户输出的样式,所述风格指定用于至少一个其他用户交互模式的表达性输出特征; 并且由多模式浏览器根据风格呈现用户输出。

    Providing Expressive User Interaction With A Multimodal Application
    2.
    发明申请
    Providing Expressive User Interaction With A Multimodal Application 有权
    提供与多模式应用程序的表达用户交互

    公开(公告)号:US20080255850A1

    公开(公告)日:2008-10-16

    申请号:US11734422

    申请日:2007-04-12

    IPC分类号: G10L21/00

    CPC分类号: G10L15/22

    摘要: Methods, apparatus, and products are disclosed for providing expressive user interaction with a multimodal application, the multimodal application operating in a multimodal browser on a multimodal device supporting multiple modes of user interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to a speech engine through a VoiceXML interpreter, including: receiving, by the multimodal browser, user input from a user through a particular mode of user interaction; determining, by the multimodal browser, user output for the user in dependence upon the user input; determining, by the multimodal browser, a style for the user output in dependence upon the user input, the style specifying expressive output characteristics for at least one other mode of user interaction; and rendering, by the multimodal browser, the user output in dependence upon the style.

    摘要翻译: 公开了方法,装置和产品,用于提供与多模式应用程序的表达性用户交互,多模式应用程序在支持多种模式的用户交互的多模式设备上的多模式浏览器中操作,包括语音模式和一种或多种非语音模式, 多模式应用通过VoiceXML解释器可操作地耦合到语音引擎,包括:由多模式浏览器通过用户交互的特定模式从用户接收用户输入; 由所述多模式浏览器根据所述用户输入确定所述用户的用户输出; 根据所述用户输入,通过所述多模式浏览器确定所述用户输出的样式,所述风格指定用于至少一个其他用户交互模式的表达性输出特征; 并且由多模式浏览器根据风格呈现用户输出。

    Altering behavior of a multimodal application based on location
    5.
    发明授权
    Altering behavior of a multimodal application based on location 有权
    基于位置改变多模式应用程序的行为

    公开(公告)号:US09208783B2

    公开(公告)日:2015-12-08

    申请号:US11679301

    申请日:2007-02-27

    CPC分类号: G10L15/22 G10L15/24

    摘要: Methods, apparatus, and products are disclosed for altering behavior of a multimodal application based on location. The multimodal application operates on a multimodal device supporting multiple modes of user interaction with the multimodal application, including a voice mode and one or more non-voice modes. The voice mode of user interaction with the multimodal application is supported by a voice interpreter. Altering behavior of a multimodal application based on location includes: receiving a location change notification in the voice interpreter from a device location manager, the device location manager operatively coupled to a position detection component of the multimodal device, the location change notification specifying a current location of the multimodal device; updating, by the voice interpreter, location-based environment parameters for the voice interpreter in dependence upon the current location of the multimodal device; and interpreting, by the voice interpreter, the multimodal application in dependence upon the location-based environment parameters.

    摘要翻译: 公开了基于位置改变多模式应用的行为的方法,装置和产品。 多模式应用程序在多模式设备上运行,支持与多模式应用程序的多种用户交互模式,包括语音模式和一种或多种非语音模式。 与多模式应用程序的用户交互的语音模式由语音解释器支持。 基于位置改变多模式应用的行为包括:从设备位置管理器在语音解释器中接收位置改变通知,该设备位置管理器可操作地耦合到多模态设备的位置检测组件,位置变化通知指定当前位置 的多模式设备; 语音解释器根据多模式设备的当前位置更新语音解释器的基于位置的环境参数; 并且由语音解释器根据基于位置的环境参数来解释多模式应用。

    Configuring a speech engine for a multimodal application based on location
    6.
    发明授权
    Configuring a speech engine for a multimodal application based on location 有权
    基于位置为多模态应用配置语音引擎

    公开(公告)号:US08938392B2

    公开(公告)日:2015-01-20

    申请号:US11679297

    申请日:2007-02-27

    IPC分类号: G10L21/00 G10L25/00 G10L15/24

    CPC分类号: G10L15/24

    摘要: Methods, apparatus, and products are disclosed for configuring a speech engine for a multimodal application based on location. The multimodal application operates on a multimodal device supporting multiple modes of user interaction with the multimodal application. The multimodal application is operatively coupled to a speech engine. Configuring a speech engine for a multimodal application based on location includes: receiving a location change notification in a location change monitor from a device location manager, the location change notification specifying a current location of the multimodal device; identifying, by the location change monitor, location-based configuration parameters for the speech engine in dependence upon the current location of the multimodal device, the location-based configuration parameters specifying a configuration for the speech engine at the current location; and updating, by the location change monitor, a current configuration for the speech engine according to the identified location-based configuration parameters.

    摘要翻译: 公开了基于位置配置用于多模式应用的语音引擎的方法,装置和产品。 多模式应用程序在支持多模式用户与多模态应用程序交互的多模式设备上运行。 多模式应用可操作地耦合到语音引擎。 基于位置为多模式应用配置语音引擎包括:从设备位置管理器在位置变化监视器中接收位置变化通知,所述位置变化通知指定多模态设备的当前位置; 根据所述多模式设备的当前位置,由所述位置变化监视器识别所述语音引擎的基于位置的配置参数,所述基于位置的配置参数指定所述语音引擎在当前位置的配置; 以及根据所识别的基于位置的配置参数,由所述位置变化监视器更新所述语音引擎的当前配置。

    SYNCHRONIZING VISUAL AND SPEECH EVENTS IN A MULTIMODAL APPLICATION
    7.
    发明申请
    SYNCHRONIZING VISUAL AND SPEECH EVENTS IN A MULTIMODAL APPLICATION 有权
    在多模式应用程序中同步视觉和语音活动

    公开(公告)号:US20120022875A1

    公开(公告)日:2012-01-26

    申请号:US13249717

    申请日:2011-09-30

    IPC分类号: G10L21/00

    CPC分类号: G10L15/1815 G10L2021/105

    摘要: Exemplary methods, systems, and products are disclosed for synchronizing visual and speech events in a multimodal application, including receiving from a user speech; determining a semantic interpretation of the speech; calling a global application update handler; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation; and executing the additional function. Typical embodiments may include updating a visual element after executing the additional function. Typical embodiments may include updating a voice form after executing the additional function. Typical embodiments also may include updating a state table after updating the voice form. Typical embodiments also may include restarting the voice form after executing the additional function.

    摘要翻译: 公开了用于在多模式应用中同步视觉和语音事件的示例性方法,系统和产品,包括从用户语音接收; 确定语音的语义解释; 调用全局应用程序更新处理程序; 由全局应用程序更新处理程序识别依赖于语义解释的附加处理功能; 并执行附加功能。 典型实施例可以包括在执行附加功能之后更新视觉元素。 典型实施例可以包括在执行附加功能之后更新语音表单。 典型实施例还可以包括在更新语音形式之后更新状态表。 典型实施例还可以包括在执行附加功能之后重新启动语音形式。

    Ordering recognition results produced by an automatic speech recognition engine for a multimodal application
    8.
    发明授权
    Ordering recognition results produced by an automatic speech recognition engine for a multimodal application 有权
    为多模式应用程序的自动语音识别引擎生成的订购识别结果

    公开(公告)号:US07840409B2

    公开(公告)日:2010-11-23

    申请号:US11679284

    申请日:2007-02-27

    IPC分类号: G10L21/06

    摘要: Ordering recognition results produced by an automatic speech recognition (‘ASR’) engine for a multimodal application implemented with a grammar of the multimodal application in the ASR engine, with the multimodal application operating in a multimodal browser on a multimodal device supporting multiple modes of interaction including a voice mode and one or more non-voice modes, the multimodal application operatively coupled to the ASR engine through a VoiceXML interpreter, includes: receiving, in the VoiceXML interpreter from the multimodal application, a voice utterance; determining, by the VoiceXML interpreter using the ASR engine, a plurality of recognition results in dependence upon the voice utterance and the grammar; determining, by the VoiceXML interpreter according to semantic interpretation scripts of the grammar, a weight for each recognition result; and sorting, by the VoiceXML interpreter, the plurality of recognition results in dependence upon the weight for each recognition result.

    摘要翻译: 通过使用ASR引擎中的多模式应用程序的语法实现的多模式应用程序的自动语音识别(“ASR”)引擎进行的订购识别结果,多模式应用程序在支持多种交互模式的多模式设备的多模式浏览器中运行 包括语音模式和一个或多个非语音模式,通过VoiceXML解释器可操作地耦合到ASR引擎的多模式应用包括:在来自多模式应用的VoiceXML解释器中接收语音话语; 通过使用ASR引擎的VoiceXML解释器,根据语音发音和语法来确定多个识别结果; 通过VoiceXML解释器根据语法的语义解释脚本确定每个识别结果的权重; 以及由VoiceXML解释器根据每个识别结果的权重对多个识别结果进行排序。

    Dynamically creating multimodal markup documents
    9.
    发明授权
    Dynamically creating multimodal markup documents 有权
    动态创建多模式标记文档

    公开(公告)号:US08032825B2

    公开(公告)日:2011-10-04

    申请号:US11154896

    申请日:2005-06-16

    IPC分类号: G06N3/00

    CPC分类号: G06F9/451

    摘要: Methods, systems, and products for dynamically creating a multimodal markup document are provided that include selecting a multimodal markup template, identifying in dependence upon the multimodal markup template a dynamic content module, instantiating the dynamic content module, executing a dynamic content creation function in the instantiated dynamic content module, receiving dynamic content from the dynamic content creation function, and including the dynamic content in the multimodal markup template. Selecting a multimodal markup template may be carried out by identifying a multimodal markup template from URI encoded data embedded in a request for a multimodal markup document from a multimodal browser. The multimodal markup template may include static content and the dynamic content may include XHTML+Voice content.

    摘要翻译: 提供了用于动态创建多模式标记文档的方法,系统和产品,其包括选择多模式标记模板,根据多模式标记模板识别动态内容模块,实例化动态内容模块,在动态内容模块中执行动态内容创建功能 实例化的动态内容模块,从动态内容创建功能接收动态内容,并将动态内容包括在多模式标记模板中。 可以通过从嵌入在多模式浏览器中的多模式标记文档的请求中的URI编码数据中识别多模式标记模板来选择多模式标记模板。 多模式标记模板可以包括静态内容,并且动态内容可以包括XHTML +语音内容。

    Synchronizing visual and speech events in a multimodal application
    10.
    发明授权
    Synchronizing visual and speech events in a multimodal application 有权
    在多模式应用程序中同步视觉和语音事件

    公开(公告)号:US08571872B2

    公开(公告)日:2013-10-29

    申请号:US13249717

    申请日:2011-09-30

    IPC分类号: G10L21/00

    CPC分类号: G10L15/1815 G10L2021/105

    摘要: Exemplary methods, systems, and products are disclosed for synchronizing visual and speech events in a multimodal application, including receiving from a user speech; determining a semantic interpretation of the speech; calling a global application update handler; identifying, by the global application update handler, an additional processing function in dependence upon the semantic interpretation; and executing the additional function. Typical embodiments may include updating a visual element after executing the additional function. Typical embodiments may include updating a voice form after executing the additional function. Typical embodiments also may include updating a state table after updating the voice form. Typical embodiments also may include restarting the voice form after executing the additional function.

    摘要翻译: 公开了用于在多模式应用中同步视觉和语音事件的示例性方法,系统和产品,包括从用户语音接收; 确定语音的语义解释; 调用全局应用程序更新处理程序; 由全局应用程序更新处理程序识别依赖于语义解释的附加处理功能; 并执行附加功能。 典型实施例可以包括在执行附加功能之后更新视觉元素。 典型实施例可以包括在执行附加功能之后更新语音表单。 典型实施例还可以包括在更新语音形式之后更新状态表。 典型实施例还可以包括在执行附加功能之后重新启动语音形式。