Online Dictionary Extension of Word Vectors
    291.
    发明申请

    公开(公告)号:US20190286716A1

    公开(公告)日:2019-09-19

    申请号:US15924791

    申请日:2018-03-19

    Applicant: Adobe Inc.

    Inventor: Zhe Lin Yingwei Li

    Abstract: Online dictionary extension of word vectors techniques and systems are described that are configured to provide online extension of existing word vector dictionaries and thus overcome the failures of conventional techniques. In one example, a dictionary extension system is employed by a computing system to extend a word vector dictionary to incorporate a new word in an online manner Co-occurrence information is estimated for the new word with respect to the words in the existing dictionary. This is done by estimating co-occurrence information with respect to a large word set based on the existing dictionary and sparse co-occurrence information for the new word. The estimated co-occurrence information is utilized to estimate a new word vector associated with the new word by projecting the estimated co-occurrence information onto the existing word vector dictionary. An extended dictionary is created incorporating the original dictionary and the estimated new word vector.

    Automatically segmenting images based on natural language phrases

    公开(公告)号:US10410351B2

    公开(公告)日:2019-09-10

    申请号:US16116609

    申请日:2018-08-29

    Applicant: Adobe Inc.

    Abstract: The invention is directed towards segmenting images based on natural language phrases. An image and an n-gram, including a sequence of tokens, are received. An encoding of image features and a sequence of token vectors are generated. A fully convolutional neural network identifies and encodes the image features. A word embedding model generates the token vectors. A recurrent neural network (RNN) iteratively updates a segmentation map based on combinations of the image feature encoding and the token vectors. The segmentation map identifies which pixels are included in an image region referenced by the n-gram. A segmented image is generated based on the segmentation map. The RNN may be a convolutional multimodal RNN. A separate RNN, such as a long short-term memory network, may iteratively update an encoding of semantic features based on the order of tokens. The first RNN may update the segmentation map based on the semantic feature encoding.

    Image Cropping Suggestion Using Multiple Saliency Maps

    公开(公告)号:US20190244327A1

    公开(公告)日:2019-08-08

    申请号:US16384593

    申请日:2019-04-15

    Applicant: Adobe Inc.

    CPC classification number: G06T3/40 G06K9/4671 G06T3/0012 G06T11/60 G06T2210/22

    Abstract: Image cropping suggestion using multiple saliency maps is described. In one or more implementations, component scores, indicative of visual characteristics established for visually-pleasing croppings, are computed for candidate image croppings using multiple different saliency maps. The visual characteristics on which a candidate image cropping is scored may be indicative of its composition quality, an extent to which it preserves content appearing in the scene, and a simplicity of its boundary. Based on the component scores, the croppings may be ranked with regard to each of the visual characteristics. The rankings may be used to cluster the candidate croppings into groups of similar croppings, such that croppings in a group are different by less than a threshold amount and croppings in different groups are different by at least the threshold amount. Based on the clustering, croppings may then be chosen, e.g., to present them to a user for selection.

    Utilizing a digital canvas to conduct a spatial-semantic search for digital visual media

    公开(公告)号:US10346727B2

    公开(公告)日:2019-07-09

    申请号:US15429769

    申请日:2017-02-10

    Applicant: Adobe Inc.

    Abstract: The present disclosure includes methods and systems for searching for digital visual media based on semantic and spatial information. In particular, one or more embodiments of the disclosed systems and methods identify digital visual media displaying targeted visual content in a targeted region based on a query term and a query area provide via a digital canvas. Specifically, the disclosed systems and methods can receive user input of a query term and a query area and provide the query term and query area to a query neural network to generate a query feature set. Moreover, the disclosed systems and methods can compare the query feature set to digital visual media feature sets. Further, based on the comparison, the disclosed systems and methods can identify digital visual media portraying targeted visual content corresponding to the query term within a targeted region corresponding to the query area.

    GENERATING MODIFIED DIGITAL IMAGES USING DEEP VISUAL GUIDED PATCH MATCH MODELS FOR IMAGE INPAINTING

    公开(公告)号:US20250139748A1

    公开(公告)日:2025-05-01

    申请号:US19011235

    申请日:2025-01-06

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately, efficiently, and flexibly generating modified digital images utilizing a guided inpainting approach that implements a patch match model informed by a deep visual guide. In particular, the disclosed systems can utilize a visual guide algorithm to automatically generate guidance maps to help identify replacement pixels for inpainting regions of digital images utilizing a patch match model. For example, the disclosed systems can generate guidance maps in the form of structure maps, depth maps, or segmentation maps that respectively indicate the structure, depth, or segmentation of different portions of digital images. Additionally, the disclosed systems can implement a patch match model to identify replacement pixels for filling regions of digital images according to the structure, depth, and/or segmentation of the digital images.

    Generating unified embeddings from multi-modal canvas inputs for image retrieval

    公开(公告)号:US12271983B2

    公开(公告)日:2025-04-08

    申请号:US17809494

    申请日:2022-06-28

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer-readable media that implements related image search and image modification processes using various search engines and a consolidated graphical user interface. For instance, in one or more embodiments, the disclosed systems receive an input digital image and search input and further modify the input digital image using the image search results retrieved in response to the search input. In some cases, the search input includes a multi-modal search input having multiple queries (e.g., an image query and a text query), and the disclosed systems retrieve the image search results utilizing a weighted combination of the queries. In some implementations, the disclosed systems generate an input embedding for the search input (e.g., the multi-modal search input) and retrieve the image search results using the input embedding.

    TEXT-AUGMENTED OBJECT CENTRIC RELATIONSHIP DETECTION

    公开(公告)号:US20250095393A1

    公开(公告)日:2025-03-20

    申请号:US18470778

    申请日:2023-09-20

    Applicant: ADOBE INC.

    Abstract: A method, apparatus, and non-transitory computer readable medium for image processing are described. Embodiments of the present disclosure obtain an image and an input text including a subject from the image and a location of the subject in the image. An image encoder encodes the image to obtain an image embedding. A text encoder encodes the input text to obtain a text embedding. An image processing apparatus based on the present disclosure generates an output text based on the image embedding and the text embedding. In some examples, the output text includes a relation of the subject to an object from the image and a location of the object in the image.

Patent Agency Ranking