LONG-TAIL COLOR PREDICTION
    41.
    发明公开

    公开(公告)号:US20240037906A1

    公开(公告)日:2024-02-01

    申请号:US17814921

    申请日:2022-07-26

    Applicant: ADOBE INC.

    CPC classification number: G06V10/764 G06V10/56 G06V10/774 G06V2201/10

    Abstract: Systems and methods for color prediction are described. Embodiments of the present disclosure receive an image that includes an object including a color, generate a color vector based on the image using a color classification network, where the color vector includes a color value corresponding to each of a set of colors, generate a bias vector by comparing the color vector to teach of a set of center vectors, where each of the set of center vectors corresponds to a color of the set of colors, and generate an unbiased color vector based on the color vector and the bias vector, where the unbiased color vector indicates the color of the object.

    Dialog system training using a simulated user system

    公开(公告)号:US11468880B2

    公开(公告)日:2022-10-11

    申请号:US16198302

    申请日:2018-11-21

    Applicant: Adobe Inc.

    Abstract: Dialog system training techniques using a simulated user system are described. In one example, a simulated user system supports multiple agents. The dialog system, for instance, may be configured for use with an application (e.g., digital image editing application). The simulated user system may therefore simulate user actions involving both the application and the dialog system which may be used to train the dialog system. Additionally, the simulated user system is not limited to simulation of user interactions by a single input mode (e.g., natural language inputs), but also supports multimodal inputs. Further, the simulated user system may also support use of multiple goals within a single dialog session

    TRAINING OF NEURAL NETWORK BASED NATURAL LANGUAGE PROCESSING MODELS USING DENSE KNOWLEDGE DISTILLATION

    公开(公告)号:US20210182662A1

    公开(公告)日:2021-06-17

    申请号:US16717698

    申请日:2019-12-17

    Applicant: Adobe Inc.

    Abstract: Techniques for training a first neural network (NN) model using a pre-trained second NN model are disclosed. In an example, training data is input to the first and second models. The training data includes masked tokens and unmasked tokens. In response, the first model generates a first prediction associated with a masked token and a second prediction associated with an unmasked token, and the second model generates a third prediction associated with the masked token and a fourth prediction associated with the unmasked token. The first model is trained, based at least in part on the first, second, third, and fourth predictions. In another example, a prediction associated with a masked token, a prediction associated with an unmasked token, and a prediction associated with whether two sentences of training data are adjacent sentences are received from each of the first and second models. The first model is trained using the predictions.

    Collecting multimodal image editing requests

    公开(公告)号:US10769495B2

    公开(公告)日:2020-09-08

    申请号:US16052246

    申请日:2018-08-01

    Applicant: Adobe Inc.

    Abstract: In implementations of collecting multimodal image editing requests (IERs), a user interface is generated that exposes an image pair including a first image and a second image including at least one edit to the first image. A user simultaneously speaks a voice command and performs a user gesture that describe an edit of the first image used to generate the second image. The user gesture and the voice command are simultaneously recorded and synchronized with timestamps. The voice command is played back, and the user transcribes their voice command based on the play back, creating an exact transcription of their voice command. Audio samples of the voice command with respective timestamps, coordinates of the user gesture with respective timestamps, and a transcription are packaged as a structured data object for use as training data to train a neural network to recognize multimodal IERs in an image editing application.

    Collecting Multimodal Image Editing Requests
    50.
    发明申请

    公开(公告)号:US20200042286A1

    公开(公告)日:2020-02-06

    申请号:US16052246

    申请日:2018-08-01

    Applicant: Adobe Inc.

    Abstract: In implementations of collecting multimodal image editing requests (IERs), a user interface is generated that exposes an image pair including a first image and a second image including at least one edit to the first image. A user simultaneously speaks a voice command and performs a user gesture that describe an edit of the first image used to generate the second image. The user gesture and the voice command are simultaneously recorded and synchronized with timestamps. The voice command is played back, and the user transcribes their voice command based on the play back, creating an exact transcription of their voice command. Audio samples of the voice command with respective timestamps, coordinates of the user gesture with respective timestamps, and a transcription are packaged as a structured data object for use as training data to train a neural network to recognize multimodal IERs in an image editing application.

Patent Agency Ranking