Personalized Automated Machine Learning

    公开(公告)号:US20210271956A1

    公开(公告)日:2021-09-02

    申请号:US16805019

    申请日:2020-02-28

    Abstract: In accordance with an embodiment of the invention, a method is provided for personalizing machine learning models for users of an automated machine learning system, the machine learning models being generated by an automated machine learning system. The method includes obtaining a first set of datasets for training first, second, and third neural networks, inputting the training datasets to the neural networks, tuning hyperparameters for the first, second, and third neural networks for testing and training the neural networks, inputting a second set of datasets to the trained neural networks and the third neural network generating a third output data including a relevance score for each of the users for each of the machine learning models, and displaying a list of machine learning models associated with each of the users, with each of the machine learning models showing the relevance score.

    Generating synchronized sound from videos

    公开(公告)号:US11039043B1

    公开(公告)日:2021-06-15

    申请号:US16744471

    申请日:2020-01-16

    Abstract: Embodiments herein describe an audio forwarding regularizer and an information bottleneck that are used when training a machine learning (ML) system. The audio forwarding regularizer receives audio training data and identifies visually irrelevant and relevant sounds in the training data. By controlling the information bottleneck, the audio forwarding regularizer forwards data to a generator that is primarily related to the visually irrelevant sounds, while filtering out the visually relevant sounds. The generator also receives data regarding visual objects from a visual encoder derived from visual training data. Thus, when being trained, the generator receives data regarding the visual objects and data regarding the visually irrelevant sounds (but little to no data regarding the visually relevant sounds). Thus, during an execution stage, the generator can generate sounds that are relevant to the visual objects while not adding visually irrelevant sounds to the videos.

    CONTEXT-AWARE CONVERSATION THREAD DETECTION FOR COMMUNICATION SESSIONS

    公开(公告)号:US20210110266A1

    公开(公告)日:2021-04-15

    申请号:US16597937

    申请日:2019-10-10

    Abstract: A computer system identifies threads in a communication session. A feature vector is generated for a message in a communication session, wherein the feature vector includes elements for features and contextual information of the message. The message feature vector and feature vectors for a plurality of threads are processed using machine learning models each associated with a corresponding thread to determine a set of probability values for classifying the message into at least one thread, wherein the threads include one or more pre-existing threads and a new thread. A classification of the message into at least one of the threads is indicated based on the set of probability values. Classification of one or more prior messages is adjusted based on the message's classification. Embodiments of the present invention further include a method and program product for identifying threads in a communication session in substantially the same manner described above.

    LEARNING DATA-AUGMENTATION FROM UNLABELED MEDIA

    公开(公告)号:US20200242507A1

    公开(公告)日:2020-07-30

    申请号:US16257965

    申请日:2019-01-25

    Abstract: A computing system is configured to learn data-augmentations from unlabeled media. The system includes an extracting unit and an embedding unit. The extracting unit is configured to receive media data that includes moving images of an object and audio generated by the object. The extracting unit extracts an image frame of the object among the moving images and extracts an audio segment from the audio. The embedding unit is configured to generate first embeddings of the image frame and second embeddings of the audio segment, and to concatenate the first and second embeddings together to generate concatenated embeddings. The computing system labels the media data based at least in part on the concatenated embeddings.

    TRAINING A POSE ESTIMATION MODEL TO DETERMINE ANATOMY KEYPOINTS IN IMAGES

    公开(公告)号:US20240404106A1

    公开(公告)日:2024-12-05

    申请号:US18327608

    申请日:2023-06-01

    Abstract: Provided are a computer program product, system, and method for training a pose estimation model to determine anatomy keypoints in images. A teacher network, implementing machine learning, processes images representing anatomies to produce heatmaps representing keypoints of the anatomies. An anatomy parsing network, implementing machine learning, processes the images to produce segmentation representations labeling anatomies represented in the images. The segmentation representations from the anatomy parsing network and the heatmaps from the teacher network are concatenated to produce mixed heatmaps. A pose estimation model, implementing machine learning, is trained to process the images to output predicted heatmaps to minimize a loss function of the output predicted heatmaps from the pose estimation model and the mixed heatmaps.

Patent Agency Ranking