Adversarially robust visual fingerprinting and image provenance models

    公开(公告)号:US12183056B2

    公开(公告)日:2024-12-31

    申请号:US17573041

    申请日:2022-01-11

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media that utilize a deep visual fingerprinting model with parameters learned from robust contrastive learning to identify matching digital images and image provenance information. For example, the disclosed systems utilize an efficient learning procedure that leverages training on bounded adversarial examples to more accurately identify digital images (including adversarial images) with a small computational overhead. To illustrate, the disclosed systems utilize a first objective function that iteratively identifies augmentations to increase contrastive loss. Moreover, the disclosed systems utilize a second objective function that iteratively learns parameters of a deep visual fingerprinting model to reduce the contrastive loss. With these learned parameters, the disclosed systems utilize the deep visual fingerprinting model to generate visual fingerprints for digital images, retrieve and match digital images, and provide digital image provenance information.

    DECENTRALIZED PLATFORM FOR ARTIFICIAL INTELLIGENCE DATA PROVENANCE AND COMMODIFICATION

    公开(公告)号:US20240281504A1

    公开(公告)日:2024-08-22

    申请号:US18172714

    申请日:2023-02-22

    Applicant: ADOBE INC.

    CPC classification number: G06F21/105 H04L9/3213 H04L2209/603

    Abstract: Systems and methods for managing rights for creative work are provided. One aspect of the systems and methods includes receiving, at a rights contract on a distributed virtual machine operated based on a public blockchain, input data including an ownership token identifier for an ownership token, where the ownership token indicates ownership of a creative work. Another aspect of the systems and methods includes obtaining, at the rights contract, an indication of usage rights for the creative work corresponding to the ownership token. Yet another aspect of the systems and methods includes minting, via the rights contract, a rights token corresponding to the ownership token, where the rights token includes a reference to the indication of the usage rights for the creative work.

    IDENTIFYING AND LOCALIZING EDITORIAL CHANGES TO IMAGES UTILIZING DEEP LEARNING

    公开(公告)号:US20230386054A1

    公开(公告)日:2023-11-30

    申请号:US17804376

    申请日:2022-05-27

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media that utilize deep learning to identify regions of an image that have been editorially modified. For example, the image comparison system includes a deep image comparator model that compares a pair of images and localizes regions that have been editorially manipulated relative to an original or trusted image. More specifically, the deep image comparator model generates and surfaces visual indications of the location of such editorial changes on the modified image. The deep image comparator model is robust and ignores discrepancies due to benign image transformations that commonly occur during electronic image distribution. The image comparison system optionally includes an image retrieval model utilizes a visual search embedding that is robust to minor manipulations or benign modifications of images. The image retrieval model utilizes a visual search embedding for an image to robustly identify near duplicate images.

    Utilizing voxel feature transformations for view synthesis

    公开(公告)号:US11823322B2

    公开(公告)日:2023-11-21

    申请号:US17807337

    申请日:2022-06-16

    Applicant: Adobe Inc.

    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for utilizing an encoder-decoder architecture to learn a volumetric 3D representation of an object using digital images of the object from multiple viewpoints to render novel views of the object. For instance, the disclosed systems can utilize patch-based image feature extraction to extract lifted feature representations from images corresponding to different viewpoints of an object. Furthermore, the disclosed systems can model view-dependent transformed feature representations using learned transformation kernels. In addition, the disclosed systems can recurrently and concurrently aggregate the transformed feature representations to generate a 3D voxel representation of the object. Furthermore, the disclosed systems can sample frustum features using the 3D voxel representation and transformation kernels. Then, the disclosed systems can utilize a patch-based neural rendering approach to render images from frustum feature patches to display a view of the object from various viewpoints.

    DETERMINING FINE-GRAIN VISUAL STYLE SIMILARITIES FOR DIGITAL IMAGES BY EXTRACTING STYLE EMBEDDINGS DISENTANGLED FROM IMAGE CONTENT

    公开(公告)号:US20220092108A1

    公开(公告)日:2022-03-24

    申请号:US17025041

    申请日:2020-09-18

    Applicant: Adobe Inc.

    Abstract: The present disclosure relates to systems, methods, and non-transitory computer readable media for accurately and flexibly identifying digital images with similar style to a query digital image using fine-grain style determination via weakly supervised style extraction neural networks. For example, the disclosed systems can extract a style embedding from a query digital image using a style extraction neural network such as a novel two-branch autoencoder architecture or a weakly supervised discriminative neural network. The disclosed systems can generate a combined style embedding by combining complementary style embeddings from different style extraction neural networks. Moreover, the disclosed systems can search a repository of digital images to identify digital images with similar style to the query digital image. The disclosed systems can also learn parameters for one or more style extraction neural network through weakly supervised training without a specifically labeled style ontology for sample digital images.

    Learning to search user experience designs based on structural similarity

    公开(公告)号:US11704559B2

    公开(公告)日:2023-07-18

    申请号:US16904460

    申请日:2020-06-17

    Applicant: Adobe Inc.

    Inventor: John Collomosse

    Abstract: Embodiments are disclosed for learning structural similarity of user experience (UX) designs using machine learning. In particular, in one or more embodiments, the disclosed systems and methods comprise generating a representation of a layout of a graphical user interface (GUI), the layout including a plurality of control components, each control component including a control type, geometric features, and relationship features to at least one other control component, generating a search embedding for the representation of the layout using a neural network, and querying a repository of layouts in embedding space using the search embedding to obtain a plurality of layouts based on similarity to the layout of the GUI in the embedding space.

    UTILIZING VOXEL FEATURE TRANSFORMATIONS FOR VIEW SYNTHESIS

    公开(公告)号:US20220327767A1

    公开(公告)日:2022-10-13

    申请号:US17807337

    申请日:2022-06-16

    Applicant: Adobe Inc.

    Abstract: Systems, methods, and non-transitory computer-readable media are disclosed for utilizing an encoder-decoder architecture to learn a volumetric 3D representation of an object using digital images of the object from multiple viewpoints to render novel views of the object. For instance, the disclosed systems can utilize patch-based image feature extraction to extract lifted feature representations from images corresponding to different viewpoints of an object. Furthermore, the disclosed systems can model view-dependent transformed feature representations using learned transformation kernels. In addition, the disclosed systems can recurrently and concurrently aggregate the transformed feature representations to generate a 3D voxel representation of the object. Furthermore, the disclosed systems can sample frustum features using the 3D voxel representation and transformation kernels. Then, the disclosed systems can utilize a patch-based neural rendering approach to render images from frustum feature patches to display a view of the object from various viewpoints.

Patent Agency Ranking