MODEL TRAINING
    32.
    发明申请

    公开(公告)号:US20220198153A1

    公开(公告)日:2022-06-23

    申请号:US17694034

    申请日:2022-03-14

    Abstract: A model training method, a model training platform, an electronic device and a storage medium are provided, which can be used in the field of artificial intelligence, particularly the fields of natural language processing and deep learning. The model training method includes: receiving an input; determining, based on the input, a user-oriented prefabricated function; determining, based on the input, a model training function; determining, based on the input, a pre-trained model; determining, based on the input, a network structure associated with the pre-trained model so as to support use of the pre-trained model; training, based on the input, the model by using the prefabricated function, the model training function, and the pre-trained model; and providing an output associated with a trained model.

    METHOD FOR PROCESSING QUERY-RESPONSE INFORMATION, METHOD FOR TRAINING MODEL, ELECTRONIC DEVICE AND MEDIUM

    公开(公告)号:US20250103963A1

    公开(公告)日:2025-03-27

    申请号:US18969634

    申请日:2024-12-05

    Abstract: A method for processing a query-response information is provided, which relates to a field of artificial intelligence technology, and in particular to fields of deep learning, large models, intelligent query and response, etc. The method for processing a query-response information includes: generating at least one initial response information according to a query information provided by an object; acquiring at least one feedback information corresponding to the at least one initial response information, wherein the feedback information indicates a preference degree of the object for the initial response information; and generating a training sample according to the query information, the at least one initial response information and the at least one feedback information. The present disclosure further provides a method for training a conversational model, an electronic device, and a storage medium.

    METHOD FOR EVALUATING LARGE MODEL, ELECTRONIC DEVICE AND COMPUTER READABLE STORAGE MEDIUM

    公开(公告)号:US20250094789A1

    公开(公告)日:2025-03-20

    申请号:US18968810

    申请日:2024-12-04

    Abstract: A method for evaluating a large model, an electronic device and a computer readable storage medium are provided, which relate to a field of artificial intelligence technology, and in particular to fields of large models technology and deep learning technology. The method includes: evaluating a response information of each of M large language models for an input instruction based on a preset evaluation rule, so as to obtain a first evaluation information for each response information, where M is a positive integer greater than 1; evaluating, in response to the first evaluation information for the M large language models being consistent with each other, each response information in a plurality of evaluation dimensions, so as to obtain a second evaluation information for each response information; and determining an evaluation result representing a responsiveness of each large language model, according to the second evaluation information for each response information.

    DIALOGUE MODEL TRAINING METHOD
    35.
    发明申请

    公开(公告)号:US20240412002A1

    公开(公告)日:2024-12-12

    申请号:US18747641

    申请日:2024-06-19

    Abstract: A method is provided. The method includes: obtaining a first sample dataset; inputting at least one first question text corresponding to at least one piece of first sample data into a dialog model separately to obtain at least one first answer prediction result; inputting each second question text into the dialog model to obtain a second answer prediction result output by the dialog model; inputting the second answer prediction result into a reward model to obtain a score of the second answer prediction result output by the reward model; determining a comprehensive loss based on the at least one first answer prediction result, a first answer text of each of the at least one piece of first sample data, and a score corresponding to each of at least one piece of second sample data; and adjusting at least one parameter of the dialog model based on the comprehensive loss.

    TEXT PROCESSING METHOD
    36.
    发明申请

    公开(公告)号:US20230101401A1

    公开(公告)日:2023-03-30

    申请号:US18056197

    申请日:2022-11-16

    Abstract: A text processing method is provided. The method includes: a first probability value of each candidate character of a plurality of candidate characters corresponding to a target position is determined based on character feature information corresponding to the target position in a text fragment to be processed, wherein the character feature information is determined based on a context at the target position in the text fragment to be processed; a second probability value of each candidate character of the plurality of candidate characters is determined based on a character string including the candidate character and at least one character in at least one position in the text fragment to be processed adjacent to the target position; and a correction character at the target position is determined based on the first probability value and the second probability value of each candidate character of the plurality of candidate characters.

    Translation Method, Apparatus and Storage Medium

    公开(公告)号:US20230095352A1

    公开(公告)日:2023-03-30

    申请号:US18074853

    申请日:2022-02-05

    Abstract: The present disclosure provides a translation method and apparatus, an electronic device, and a non-transitory storage medium. An implementation includes: determining an encoded feature of a sentence to be translated by an encoding module; determining, by a graph network module, a knowledge fusion feature of the sentence to be translated based on a preset graph network, wherein the preset graph network is constructed based on a polysemous word in a source language corresponding to the sentence to be translated and a plurality of translated words corresponding to the polysemous word in a target language; determining, by a decoding network, a translated sentence corresponding to the sentence to be translated based on the encoded feature and the knowledge fusion feature.

    METHOD FOR CORRECTING TEXT, METHOD FOR GENERATING TEXT CORRECTION MODEL, DEVICE

    公开(公告)号:US20230090625A1

    公开(公告)日:2023-03-23

    申请号:US18053034

    申请日:2022-11-07

    Abstract: Disclosed are a method for correcting a text, an electronic device and a storage medium. The method includes: acquiring a text to be corrected; acquiring a phonetic symbol sequence of the text to be corrected; and obtaining a corrected text by inputting the text to be corrected and the phonetic symbol sequence into a text correction model, in which, the text correction model obtains the corrected text by: detecting an error word in the text to be corrected, determining a phonetic symbol corresponding to the error word in the phonetic symbol sequence, and adding the phonetic feature corresponding to the phonetic symbol behind the error word to obtain a phonetic symbol text, and correcting the error word and the phonetic feature in the phonetic symbol text to obtain the corrected text.

    METHOD OF TRAINING DEEP LEARNING MODEL AND METHOD OF PROCESSING TEXT DATA

    公开(公告)号:US20230088360A1

    公开(公告)日:2023-03-23

    申请号:US18059389

    申请日:2022-11-28

    Abstract: A method of training a deep learning model is provided, which relates to a field of artificial intelligence, in particular to a field of a natural language processing technology and a field of a machine translation technology. A specific implementation solution includes: processing sample source data and corresponding sample target data respectively by using the deep learning model, so as to obtain a first output value and a second output value; determining a regularization function value according to the first output value and the second output value; and adjusting a parameter of the depth learning model according to the regularization function value, so as to obtain a pre-trained depth learning model. A method of processing text data, an electronic device, and a storage medium are further provided.

Patent Agency Ranking