PROMPT TUNING FOR ZERO-SHOT COMPOSITIONAL LEARNING IN MACHINE LEARNING SYSTEMS

    公开(公告)号:US20240203143A1

    公开(公告)日:2024-06-20

    申请号:US18454459

    申请日:2023-08-23

    CPC classification number: G06V20/70 G06F40/284 G06V10/774

    Abstract: A method includes obtaining an image, a set of attribute labels, and a set of object labels and performing prompt tuning of a pre-trained vision-language model having first and second textual encoders and a vision encoder. The model is trained during prompt tuning to select one attribute label and one object label that match content in the image. Performing the prompt tuning includes, for each attribute label-object label pair, generating object textual features associated with the object label using the first textual encoder, generating attribute textual features associated with the attribute label using the second textual encoder, and generating image features associated with the image using the vision encoder. Intermediate outputs from initial layers of the textual encoders and the vision encoder are combined to generate layer-specific learnable prompt tokens that are appended to inputs of specified layers in the first and second textual encoders and the vision encoder.

    System and method for deep memory network

    公开(公告)号:US11775815B2

    公开(公告)日:2023-10-03

    申请号:US16535380

    申请日:2019-08-08

    CPC classification number: G06N3/08 G06N5/04

    Abstract: An electronic device including a deep memory model includes at least one memory and at least one processor coupled to the at least one memory. The at least one processor is configured to receive input data to the deep memory model. The at least one processor is also configured to extract a history state of an external memory coupled to the deep memory model based on the input data. The at least one processor is further configured to update the history state of the external memory based on the input data. In addition, the at least one processor is configured to output a prediction based on the extracted history state of the external memory.

    System and method for explaining and compressing deep learning natural language understanding (NLU) models

    公开(公告)号:US11455471B2

    公开(公告)日:2022-09-27

    申请号:US16947258

    申请日:2020-07-24

    Abstract: A method includes obtaining, using at least one processor of an electronic device, a base natural language understanding (NLU) model that includes a word embedding layer, where the word embedding layer is associated with at least one training utterance. The method also includes calculating, using the at least one processor, a regularization loss value for use in a determination of an intent detection loss, where the regularization loss value reveals an effect of word embeddings on intent determination of the training utterance. The method further includes retraining, using the at least one processor, the word embedding layer of the base NLU model using the intent detection loss to obtain a retrained NLU model.

    SYSTEMS AND METHODS FOR CONTINUAL LEARNING

    公开(公告)号:US20210383272A1

    公开(公告)日:2021-12-09

    申请号:US17166908

    申请日:2021-02-03

    Abstract: A continual learning method includes obtaining an input data including a trained model, continual learning (CL) Information, and training data by an electronic device. The method also includes re-training, using the electronic device, the model for a task based on the training data. The method also includes updating, using the electronic device, the CL Information based on the model and the training data. The method further includes selecting a first set of exemplars from the training data based on data associated with the CL Information. The CL Information includes a first group of variables associated with the model and a second group of variables associated with the model that changes to the first group of variables have stronger impact to the model's performance of the task than changes to the second group of variables.

Patent Agency Ranking