JOINT MODEL TRAINING
    1.
    发明申请

    公开(公告)号:US20170132528A1

    公开(公告)日:2017-05-11

    申请号:US15195894

    申请日:2016-06-28

    IPC分类号: G06N99/00 G06N7/00

    CPC分类号: G06N20/00

    摘要: Multiple machine learning models can be jointly trained in parallel. An example process for jointly training multiple machine learning models includes providing a set of machine learning models that are to learn a respective task, the set of machine learning models including a first machine learning model and a second machine learning model. The process can initiate training of the first machine learning model to learn a task using training data. During the training of the first machine learning model, information can be passed between the first machine learning model and the second machine learning model. Such passing of information (or “transfer of knowledge”) between the machine learning models can be accomplished via the formulation, and optimization, of an objective function that comprises model parameters that are based on the multiple machine learning models in the set.

    Wearable audio accessories for computing devices

    公开(公告)号:US10264346B2

    公开(公告)日:2019-04-16

    申请号:US15782185

    申请日:2017-10-12

    IPC分类号: G06F3/16 H04R1/10 H04M1/60

    摘要: Wearable audio accessories for computing devices are described. In one embodiment the wearable audio accessory provides a speech based interface between the user and a nearby computing device for the performance of user-initiated or computing device initiated microtasks. Information is provided to the user via a loudspeaker and the user can provide input via a microphone. An audio sensing channel within the accessory continuously monitors the audio signal as detected by the microphone and in various embodiments will trigger more complex audio processing based on this monitoring. A wireless communication link is provided between the accessory and the nearby computing device. To mitigate any delay caused by the switching between audio processing techniques, the audio accessory may include a rolling buffer which continuously stores the audio signal and outputs a delayed audio signal to the audio processing engines.

    Live video analytics at scale
    3.
    发明授权

    公开(公告)号:US10223604B2

    公开(公告)日:2019-03-05

    申请号:US15373301

    申请日:2016-12-08

    摘要: Various technologies described herein pertain to performing video analytics. The approaches set forth herein support live video analytics at scale with approximate and delay-tolerant processing. Video streams can be captured by multiple cameras and continuously streamed to a video analytics computing system; the video streams can be received at the video analytics computing system. Multiple video analytics queries can be executed on the video streams. The multiple video analytics queries can be concurrently executed by the video analytics computing system on the video streams as the video streams are continuously streamed to the video analytics computing system. The multiple video analytics queries can be executed utilizing resources of the video analytics computing system allocated between the multiple video analytics queries. Execution of the multiple video analytics queries can return respective results for the multiple video analytics queries. The results for the multiple video analytics queries can be outputted.

    Querying video data with reduced latency and cost

    公开(公告)号:US11354902B2

    公开(公告)日:2022-06-07

    申请号:US16875080

    申请日:2020-05-15

    IPC分类号: G06V20/40 G06K9/62

    摘要: A method can include classifying, using a compressed and specialized convolutional neural network (CNN), an object of a video frame into classes, clustering the object based on a distance of a feature vector of the object to a feature vector of a centroid object of the cluster, storing top-k classes, a centroid identification, and a cluster identification, in response to receiving a query for objects of class X from a specific video stream, retrieving image data for each centroid of each cluster that includes the class X as one of the top-k classes, classifying, using a ground truth CNN (GT-CNN), the retrieved image data for each centroid, and for each centroid determined to be classified as a member of the class X providing image data for each object in each cluster associated with the centroid.

    PERSONAL TRANSLATOR
    9.
    发明申请
    PERSONAL TRANSLATOR 审中-公开
    个人翻译

    公开(公告)号:US20170060850A1

    公开(公告)日:2017-03-02

    申请号:US14834197

    申请日:2015-08-24

    摘要: The personal translator implementations described herein provide a speech translation device that pairs with a computing device to translate in-person conversations. The speech translation device can be wearable. In one implementation the personal translator comprises a speech translation device with at least one microphone that captures input signals representing nearby speech of a first user/wearer of the device and at least one other nearby person in a conversation in two languages; a wireless communication unit that sends the captured input signals representing speech to a nearby computing device, and receives for each language in the conversation, language translations from the computing device; and at least one loudspeaker that outputs the language translations to the first user/wearer and at least one other nearby person. The language translations in text form can be displayed on a display at the same time the language translations are output to the loudspeaker(s).

    摘要翻译: 本文描述的个人翻译器实现提供了与计算设备配对以翻译个人对话的语音翻译设备。 语音翻译装置可以穿戴。 在一个实施方式中,个人翻译器包括具有至少一个麦克风的语音翻译装置,该麦克风在两种语言的对话中捕获表示设备的第一用户/穿戴者的附近语音的输入信号和至少一个其他附近的人; 无线通信单元,其将表示语音的所捕获的输入信号发送到附近的计算设备,并且接收对话中的每种语言,来自所述计算设备的语言翻译; 以及至少一个扬声器,其将语言翻译输出到第一用户/佩戴者以及至少一个其他附近的人。 文本形式的语言翻译可以在显示器上显示,同时将语言翻译输出到扬声器。

    Dynamic video highlight
    10.
    发明授权

    公开(公告)号:US11170819B2

    公开(公告)日:2021-11-09

    申请号:US16411611

    申请日:2019-05-14

    摘要: Described herein is a mechanism for creating a dynamic video highlight from a plurality of video segments. A metadata collection agent collects metadata comprising attributes about a video, segments within the video where one or more events occur, attributes about the creator of the video and so forth. The metadata is collected and used to create highlight video definitions comprising a set of metadata attribute-value pairs. The highlight video definitions can be created in an interactive fashion by presenting a user interface allowing selection of a combination of attribute-value pairs to include/exclude segments from the definition and/or manual selection of custom segments. The highlight video definitions can be stored and/or shared among users. The video highlight definitions are utilized to instantiate one or more video players to play the video segments in an identified order without assembling the video segments into a separate video.