SEMANTIC IMAGE MANIPULATION USING VISUAL-SEMANTIC JOINT EMBEDDINGS

    公开(公告)号:US20220036127A1

    公开(公告)日:2022-02-03

    申请号:US16943511

    申请日:2020-07-30

    Applicant: Adobe Inc.

    Abstract: The technology described herein is directed to a reinforcement learning based framework for training a natural media agent to learn a rendering policy without human supervision or labeled datasets. The reinforcement learning based framework feeds the natural media agent a training dataset to implicitly learn the rendering policy by exploring a canvas and minimizing a loss function. Once trained, the natural media agent can be applied to any reference image to generate a series (or sequence) of continuous-valued primitive graphic actions, e.g., sequence of painting strokes, that when rendered by a synthetic rendering environment on a canvas, reproduce an identical or transformed version of the reference image subject to limitations of an action space and the learned rendering policy.

    Semantic image manipulation using visual-semantic joint embeddings

    公开(公告)号:US11574142B2

    公开(公告)日:2023-02-07

    申请号:US16943511

    申请日:2020-07-30

    Applicant: Adobe Inc.

    Abstract: The technology described herein is directed to a reinforcement learning based framework for training a natural media agent to learn a rendering policy without human supervision or labeled datasets. The reinforcement learning based framework feeds the natural media agent a training dataset to implicitly learn the rendering policy by exploring a canvas and minimizing a loss function. Once trained, the natural media agent can be applied to any reference image to generate a series (or sequence) of continuous-valued primitive graphic actions, e.g., sequence of painting strokes, that when rendered by a synthetic rendering environment on a canvas, reproduce an identical or transformed version of the reference image subject to limitations of an action space and the learned rendering policy.

    Joint Visual-Semantic Embedding and Grounding via Multi-Task Training for Image Searching

    公开(公告)号:US20210271707A1

    公开(公告)日:2021-09-02

    申请号:US16803480

    申请日:2020-02-27

    Applicant: Adobe Inc.

    Abstract: Certain embodiments involve a method for generating a search result. The method includes processing devices performing operations including receiving a query having a text input by a joint embedding model trained to generate an image result. Training the joint embedding model includes accessing a set of images and textual information. Training further includes encoding the images into image feature vectors based on spatial features. Further, training includes encoding the textual information into textual feature vectors based on semantic information. Training further includes generating a set of image-text pairs based on matches between image feature vectors and textual feature vectors. Further, training includes generating a visual grounding dataset based on spatial information. Training further includes generating a set of visual-semantic joint embeddings by grounding the image-text pairs with the visual grounding dataset. Additionally, operations include generating an image result for display by the joint embedding model based on the text input.

Patent Agency Ranking