A PEPTIDE BASED VACCINE GENERATION SYSTEM WITH DUAL PROJECTION GENERATIVE ADVERSARIAL NETWORKS

    公开(公告)号:US20240029822A1

    公开(公告)日:2024-01-25

    申请号:US18479416

    申请日:2023-10-02

    CPC classification number: G16B15/30 G06N3/08 G16B40/20 G06N3/045

    Abstract: A computer-implemented method is provided for generating new binding peptides to Major Histocompatibility Complex (MHC) proteins. The method includes training, by a processor device, a Generative Adversarial Network GAN having a generator and a discriminator only on a set of binding peptide sequences given training data comprising the set of binding peptide sequences and a set of non-binding peptide sequences. A GAN training objective includes the discriminator being iteratively updated to distinguish generated peptide sequences from sampled binding peptide sequences as fake or real and the generator being iteratively updated to fool the discriminator. The training includes optimizing the GAN training objective while learning two projection vectors for a binding class with two cross-entropy losses. A first loss discriminating binding peptide sequences in the training data from non-binding peptide sequences in the training data. A second loss discriminating generated binding peptide sequences from non-binding peptide sequences in the training data.

    PEPTIDE BASED VACCINE GENERATION SYSTEM WITH DUAL PROJECTION GENERATIVE ADVERSARIAL NETWORKS

    公开(公告)号:US20240029821A1

    公开(公告)日:2024-01-25

    申请号:US18479409

    申请日:2023-10-02

    CPC classification number: G16B15/30 G06N3/08 G16B40/20 G06N3/045

    Abstract: A computer-implemented method is provided for generating new binding peptides to Major Histocompatibility Complex (MHC) proteins. The method includes training, by a processor device, a Generative Adversarial Network GAN having a generator and a discriminator only on a set of binding peptide sequences given training data comprising the set of binding peptide sequences and a set of non-binding peptide sequences. A GAN training objective includes the discriminator being iteratively updated to distinguish generated peptide sequences from sampled binding peptide sequences as fake or real and the generator being iteratively updated to fool the discriminator. The training includes optimizing the GAN training objective while learning two projection vectors for a binding class with two cross-entropy losses. A first loss discriminating binding peptide sequences in the training data from non-binding peptide sequences in the training data. A second loss discriminating generated binding peptide sequences from non-binding peptide sequences in the training data.

    SOURCE-FREE CROSS DOMAIN DETECTION METHOD WITH STRONG DATA AUGMENTATION AND SELF-TRAINED MEAN TEACHER MODELING

    公开(公告)号:US20230154167A1

    公开(公告)日:2023-05-18

    申请号:US17966017

    申请日:2022-10-14

    CPC classification number: G06V10/7747 G06V10/25 G06V10/765 G06V2201/07

    Abstract: A method for implementing source-free domain adaptive detection is presented. The method includes, in a pretraining phase, applying strong data augmentation to labeled source images to produce perturbed labeled source images and training an object detection model by using the perturbed labeled source images to generate a source-only model. The method further includes, in an adaptation phase, training a self-trained mean teacher model by generating a weakly augmented image and multiple strongly augmented images from unlabeled target images, generating a plurality of region proposals from the weakly augmented image, selecting a region proposal from the plurality of region proposals as a pseudo ground truth, detecting, by the self-trained mean teacher model, object boxes and selecting pseudo ground truth boxes by employing a confidence constraint and a consistency constraint, and training a student model by using one of the multiple strongly augmented images jointly with an object detection loss.

    COMPOSITIONAL TEXT-TO-IMAGE SYNTHESIS WITH PRETRAINED MODELS

    公开(公告)号:US20230153606A1

    公开(公告)日:2023-05-18

    申请号:US17968923

    申请日:2022-10-19

    Abstract: A method is provided that includes training a CLIP model to learn embeddings of images and text from matched image-text pairs. The text represents image attributes. The method trains a StyleGAN on images in a training dataset of matched image-text pairs. The method also trains, using a CLIP model guided contrastive loss which attracts matched text embedding pairs and repels unmatched pairs, a text-to-direction model to predict a text direction that is semantically aligned with an input text responsive to the input text and a random latent code. A triplet loss is used to learn text directions using the embeddings learned by the trained CLIP model. The method generates, by the trained StyleGAN, positive and negative synthesized images by respectively adding and subtracting the text direction in the latent space of the trained StyleGAN corresponding to a word for each of the words in the training dataset.

    PEPTIDE BASED VACCINE GENERATION SYSTEM WITH DUAL PROJECTION GENERATIVE ADVERSARIAL NETWORKS

    公开(公告)号:US20220328127A1

    公开(公告)日:2022-10-13

    申请号:US17711310

    申请日:2022-04-01

    Abstract: A computer-implemented method is provided for generating new binding peptides to Major Histocompatibility Complex (MHC) proteins. The method includes training, by a processor device, a Generative Adversarial Network GAN having a generator and a discriminator only on a set of binding peptide sequences given training data comprising the set of binding peptide sequences and a set of non-binding peptide sequences. A GAN training objective includes the discriminator being iteratively updated to distinguish generated peptide sequences from sampled binding peptide sequences as fake or real and the generator being iteratively updated to fool the discriminator. The training includes optimizing the GAN training objective while learning two projection vectors for a binding class with two cross-entropy losses. A first loss discriminating binding peptide sequences in the training data from non-binding peptide sequences in the training data. A second loss discriminating generated binding peptide sequences from non-binding peptide sequences in the training data.

    Self-supervised sequential variational autoencoder for disentangled data generation

    公开(公告)号:US11423655B2

    公开(公告)日:2022-08-23

    申请号:US17088043

    申请日:2020-11-03

    Abstract: A computer-implemented method is provided for disentangled data generation. The method includes accessing, by a variational autoencoder, a plurality of supervision signals. The method further includes accessing, by the variational autoencoder, a plurality of auxiliary tasks that utilize the supervision signals as reward signals to learn a disentangled representation. The method also includes training the variational autoencoder to disentangle a sequential data input into a time-invariant factor and a time-varying factor using a self-supervised training approach which is based on outputs of the auxiliary tasks obtained by using the supervision signals to accomplish the plurality of auxiliary tasks.

    MULTI-HOP TRANSFORMER FOR SPATIO-TEMPORAL REASONING AND LOCALIZATION

    公开(公告)号:US20220101007A1

    公开(公告)日:2022-03-31

    申请号:US17463757

    申请日:2021-09-01

    Abstract: A method for using a multi-hop reasoning framework to perform multi-step compositional long-term reasoning is presented. The method includes extracting feature maps and frame-level representations from a video stream by using a convolutional neural network (CNN), performing object representation learning and detection, linking objects through time via tracking to generate object tracks and image feature tracks, feeding the object tracks and the image feature tracks to a multi-hop transformer that hops over frames in the video stream while concurrently attending to one or more of the objects in the video stream until the multi-hop transformer arrives at a correct answer, and employing video representation learning and recognition from the objects and image context to locate a target object within the video stream.

    Network reparameterization for new class categorization

    公开(公告)号:US11087184B2

    公开(公告)日:2021-08-10

    申请号:US16580199

    申请日:2019-09-24

    Abstract: A computer-implemented method and system are provided for training a model for New Class Categorization (NCC) of a test image. The method includes decoupling, by a hardware processor, a feature extraction part from a classifier part of a deep classification model by reparametrizing learnable weight variables of the classifier part as a combination of learnable variables of the feature extraction part and of a classification weight generator of the classifier part. The method further includes training, by the hardware processor, the deep classification model to obtain a trained deep classification model by (i) learning the feature extraction part as a multiclass classification task, and (ii) episodically training the classifier part by learning a classification weight generator which outputs classification weights given a training image.

Patent Agency Ranking