Cross-Modal Contrastive Learning for Text-to-Image Generation based on Machine Learning Models

    公开(公告)号:US20230081171A1

    公开(公告)日:2023-03-16

    申请号:US17467628

    申请日:2021-09-07

    Applicant: Google LLC

    Abstract: A computer-implemented method includes receiving, by a computing device, a particular textual description of a scene. The method also includes applying a neural network for text-to-image generation to generate an output image rendition of the scene, the neural network having been trained to cause two image renditions associated with a same textual description to attract each other and two image renditions associated with different textual descriptions to repel each other based on mutual information between a plurality of corresponding pairs, wherein the plurality of corresponding pairs comprise an image-to-image pair and a text-to-image pair. The method further includes predicting the output image rendition of the scene.

    Systems And Methods For Generating Predicted Visual Observations Of An Environment Using Machine Learned Models

    公开(公告)号:US20230072293A1

    公开(公告)日:2023-03-09

    申请号:US17409249

    申请日:2021-08-23

    Applicant: Google LLC

    Abstract: A computing system for generating predicted images along a trajectory of unseen viewpoints. The system can obtain one or more spatial observations of an environment that may be captured from one or more previous camera poses. The system can generate a three-dimensional point cloud for the environment from the one or more spatial observations and the one or more previous camera poses. The system can project the three-dimensional point cloud into two-dimensional space to form one or more guidance spatial observations. The system can process the one or more guidance spatial observations with a machine-learned spatial observation prediction model to generate one or more predicted spatial observations. The system can process the one or more predicted spatial observations and image data with a machine-learned image prediction model to generate one or more predicted images from the target camera pose. The system can output the one or more predicted images.

    Vector-Quantized Image Modeling
    7.
    发明公开

    公开(公告)号:US20240112088A1

    公开(公告)日:2024-04-04

    申请号:US18520083

    申请日:2023-11-27

    Applicant: Google LLC

    CPC classification number: G06N20/00

    Abstract: Systems and methods are provided for vector-quantized image modeling using vision transformers and improved codebook handling. In particular, the present disclosure provides a Vector-quantized Image Modeling (VIM) approach that involves pretraining a machine learning model (e.g., Transformer model) to predict rasterized image tokens autoregressively. The discrete image tokens can be encoded from a learned Vision-Transformer-based VQGAN (example implementations of which can be referred to as ViT-VQGAN). The present disclosure proposes multiple improvements over vanilla VQGAN from architecture to codebook learning, yielding better efficiency and reconstruction fidelity. The improved ViT-VQGAN further improves vector-quantized image modeling tasks, including unconditional image generation, conditioned image generation (e.g., class-conditioned image generation), and unsupervised representation learning.

Patent Agency Ranking