-
21.
公开(公告)号:US20250094460A1
公开(公告)日:2025-03-20
申请号:US18969597
申请日:2024-12-05
Inventor: Haifeng WANG , Hua WU , Hao TIAN , Jing LIU , Ruiqing ZHANG , Yan CHEN , Yu RAN
IPC: G06F16/3329 , G06F16/3332 , G06F16/334
Abstract: A query answering method, an electronic device, a storage medium, and an intelligent agent are provided, which relate to a field of artificial intelligence technology, and in particular to fields of large model, intelligent search and information processing technology. The method includes: inputting, in response to a retrieval content set retrieved based on a query, the query, the retrieval content set and prompt information for answer generation into the large model, so that the large model performs operations of: processing, based on a current task in the prompt information and the query, a current text corresponding to the retrieval content set to obtain a processed text, where the current task is determined based on a task execution order in the prompt information; and obtaining, in a case of determining that the processed text meets a preset condition, an answer to the query based on the processed text.
-
公开(公告)号:US20250061305A1
公开(公告)日:2025-02-20
申请号:US18936686
申请日:2024-11-04
Inventor: Shuohuan WANG , Junyuan SHANG , Yinqi YANG , Guoxia WANG , Linhao ZHANG , Yu SUN , Hua WU , Haifeng WANG
IPC: G06N3/043 , G06N3/045 , G06N3/0985
Abstract: A training method, an inference method, a device, an apparatus, and a medium for a deep learning model are provided. A first model includes a plurality of first parameters, a second model comprises a plurality of second parameters, which is initialized to parameter values of a plurality of target parameters selected from the plurality of first parameters. The training method includes: determining a target loss for both the first model and the second model; adjusting parameter values, including: in response to determining that the target loss indicates that the parameter values of at least part of the target parameters need to be adjusted, synchronously adjusting the parameter values of the corresponding second parameters; and in response to determining that the target loss indicates that the parameter values of at least part of the second parameters need to be adjusted, synchronously adjusting the parameter values of the corresponding target parameters.
-
23.
公开(公告)号:US20240354658A1
公开(公告)日:2024-10-24
申请号:US18745529
申请日:2024-06-17
Inventor: Feng HE , Jianhua WANG , Junjie OU , Pingxuan HUANG , Zhifan FENG , Xiaopeng CUI , Qiaoqiao SHE , Hua WU
Abstract: A method and apparatus for training a question solving model, a question solving method and apparatus, an electronic device and a readable storage medium are disclosed. The method for training a question solving model includes: acquiring a first sample question; inputting the first sample question and a solving step grabbing template into a large language model to obtain a first sample solving step; inputting the first sample question, the first sample solving step and an answer grabbing template into the large language model to obtain a first sample answer; pre-training a step planning model according to the first sample question and the first sample solving step; pre-training the large language model according to the first sample question, the first sample solving step and the first sample answer; and acquiring the question solving model according to the step planning model and the large language model obtained by pre-training. The question solving method includes: acquiring a to-be-solved question; inputting the to-be-solved question into a step planning model to obtain a solving step; and inputting the to-be-solved question and the solving step into a large language model to obtain an answer.
-
公开(公告)号:US20240338862A1
公开(公告)日:2024-10-10
申请号:US18749461
申请日:2024-06-20
Inventor: Jiachen LIU , Xinyan XIAO , Hua WU , Guohao LI , Wei LI , Hong ZHU , Qiaoqiao SHE , Yajuan LV
Abstract: A method is provided that includes: obtaining current dialogue data; determining a requirement type of the user in the current round of dialogue based on the current dialogue data; in response to the requirement type being an image processing requirement, determining an action sequence for implementing the image processing requirement; executing the action sequence to generate a target image; and generating response data corresponding to the user input data based on the target image.
-
公开(公告)号:US20230342561A1
公开(公告)日:2023-10-26
申请号:US18122316
申请日:2023-03-16
Inventor: Ruiqing ZHANG , Hui LIU , Zhongjun HE , Zhi LI , Hua WU
Abstract: A machine translation method includes: obtaining first target language text by performing first translation on source language text using an initial NMT model; identifying an untranslated part in the source language text based on the source language text and the first target language text; obtaining an adjusted NMT model by increasing an attention weight corresponding to the untranslated part in the initial NMT mode; and obtaining second target language text by performing second translation on the source language text using the adjusted NMT model.
-
公开(公告)号:US20230153543A1
公开(公告)日:2023-05-18
申请号:US17951216
申请日:2022-09-23
Inventor: Ruiqing ZHANG , Xiyang WANG , Hui LIU , Zhongjun HE , Zhi LI , Hua WU
Abstract: A translation method, a model training method, apparatuses, electronic devices and storage mediums, which relate to the field of artificial intelligence technologies, such as machine learning technologies, information processing technologies, are disclosed. In an implementation, a weight for each translation model in at least two pre-trained translation models translating a to-be-translated specified sentence is acquired based on the specified sentence and a pre-trained weighting model; and the specified sentence is translating using the at least two translation models based on the weight for each translation model translating the specified sentence.
-
公开(公告)号:US20230051373A1
公开(公告)日:2023-02-16
申请号:US17974317
申请日:2022-10-26
Inventor: Xiyang WANG , Ruiqing ZHANG , Zhongjun HE , Zhi LI , Hua WU
IPC: G06F40/47
Abstract: A method for training a non-autoregressive translation (NAT) model includes: acquiring a source language text, a target language text corresponding to the source language text and a target length of the target language text; generating a target language prediction text and a prediction length by inputting the source language text into the NAT model, in which initialization parameters of the NAT model are determined based on parameters of a pre-trained translation model; and obtaining a target NAT model by training the NAT model based on the target language text, the target language prediction text, the target length and the prediction length.
-
公开(公告)号:US20230032324A1
公开(公告)日:2023-02-02
申请号:US17966127
申请日:2022-10-14
Inventor: Fan WANG , Hao TIAN , Haoyi XIONG , Hua WU , Jingzhou HE , Haifeng WANG
IPC: G06N20/00
Abstract: A method for training a decision-making model parameter, a decision determination method, an electronic device, and a non-transitory computer-readable storage medium are provided. In the method, a perturbation parameter is generated according to a meta-parameter, and first observation information of a primary training environment is acquired based on the perturbation parameter. According to the first observation information, an evaluation parameter of the perturbation parameter is determined. According to the perturbation parameter and the evaluation parameter thereof, an updated meta-parameter is generated. The updated meta-parameter is determined as a target meta-parameter, when it is determined, according to the meta-parameter and the updated meta-parameter, that a condition for stopping primary training is met. According to the target meta-parameter, a target memory parameter corresponding to a secondary training task is determined, where the target memory parameter and the target meta-parameter are used to make a decision corresponding to a prediction task.
-
公开(公告)号:US20220391602A1
公开(公告)日:2022-12-08
申请号:US17820765
申请日:2022-08-18
Inventor: Haifeng WANG , Zhanyi LIU , Zhongjun HE , Hua WU , Zhi LI , Xing WAN , Jingxuan ZHAO , Ruiqing ZHANG , Chuanqiang ZHANG , Fengtao HUANG , Hanbing SONG , Wei DI , Shuangshuang CUI , Yongzheng XIN
Abstract: A display method, an electronic device, and a storage medium, which relate to a field of natural language processing and a field of display. The display method includes: acquiring a content to be displayed; extracting a target term from the content using a term extraction rule; acquiring an annotation information for at least one target term, responsive to an extraction of the at least one target term; and displaying the annotation information for the at least one target term and the content.
-
公开(公告)号:US20220215899A1
公开(公告)日:2022-07-07
申请号:US17557691
申请日:2021-12-21
Inventor: Fan WANG , Jingzhou HE , Xiaomin FANG , Xiaonan ZHANG , Hua WU , Tian WU , Haifeng WANG
Abstract: The present disclosure discloses an affinity prediction method and apparatus, a method and apparatus for training an affinity prediction model, a device and a medium, and relates to the field of artificial intelligence technologies, such as machine learning technologies, smart medical technologies, or the like. An implementation includes: collecting a plurality of training samples, each training sample including information of a training target, information of a training drug and a test data set corresponding to the training target; and training an affinity prediction model using the plurality of training samples. In addition, there is further disclosed the affinity prediction method. The technology in the present disclosure may effectively improve accuracy and a training effect of the trained affinity prediction model. During an affinity prediction, accuracy of a predicted affinity of a target to be detected with a drug to be detected may be higher by acquiring a test data set corresponding to the target to be detected to participate in the prediction.
-
-
-
-
-
-
-
-
-