-
公开(公告)号:US20210374349A1
公开(公告)日:2021-12-02
申请号:US17444693
申请日:2021-08-09
Inventor: Jiachen LIU , Xinyan XIAO , Hua WU , Haifeng WANG
IPC: G06F40/295 , G06N20/00 , G06N5/02
Abstract: A method for text generation, relates to a field of natural language processing, including: obtaining corpus data; labeling the corpus data to obtain a first constraint element; obtaining a first generation target; and generating a first text matching the first generation target by inputting the corpus data and the first constraint element into a generation model.
-
公开(公告)号:US20210286934A1
公开(公告)日:2021-09-16
申请号:US17331526
申请日:2021-05-26
Inventor: Jiachen LIU , Zhe HU , Xinyan XIAO , Hua WU
IPC: G06F40/117 , G06F40/126 , G06N20/00
Abstract: A method for implementing text generation, a device and a medium are provided. The method includes: determining a target task type of a target text generation task from multiple task types supported by a pre-trained general text generation model; determining, based on a requirement of the target text generation task for a target output text, a first target output text attribute for the target text generation task from multiple output text attributes supported by the general text generation model; and fine tuning the general text generation model based on a target training data set associated with the target text generation task to obtain a task-specific text generation model, by taking task indication information for the target task type and first attribute indication information for the first target output text attribute as at least part of an input of the general text generation model.
-
3.
公开(公告)号:US20190163714A1
公开(公告)日:2019-05-30
申请号:US16313195
申请日:2016-09-05
Inventor: Yanjun MA , Jiachen LIU , Hua WU
IPC: G06F16/9535 , G06F16/2458 , G06F16/9538 , G06N20/00
Abstract: The present disclosure provides a search result aggregation method and apparatus based on artificial intelligence and a search engine. The method includes: obtaining a query; generating a plurality of search results according to the query; obtaining a plurality of corresponding demand dimensions according to the query; aggregating the plurality of demand dimensions according to the plurality of search results; obtaining an answer corresponding to each demand dimension, and aggregating the answers corresponding to the plurality of demand dimensions according to the aggregated demand dimensions to generate an aggregation result.
-
公开(公告)号:US20210374359A1
公开(公告)日:2021-12-02
申请号:US17133381
申请日:2020-12-23
Inventor: Wei LI , Xinyan XIAO , Hua WU , Haifeng WANG
Abstract: The disclosure may provide a method for obtaining a document layout, an electronic device, and a storage medium. The method may include: obtaining a plurality of pieces of first sample data; extracting structured information from each of the plurality of pieces of first sample data as target structured information corresponding to each of the plurality of pieces of first sample data; inputting the plurality of pieces of first sample data into an initial text generation model to generate predicted structured information corresponding to each of the plurality of pieces of first sample data; generating a first loss value based on a difference between the predicted structured information corresponding to each of the plurality of pieces of first sample data and the corresponding target structured information; and training a phrase generation ability of the initial text generation model based on the first loss value to generate the text generation model.
-
公开(公告)号:US20210200957A1
公开(公告)日:2021-07-01
申请号:US16895297
申请日:2020-06-08
Inventor: Siqi BAO , Huang HE , Junkun CHEN , Fan WANG , Hua WU , Jingzhou HE
IPC: G06F40/30
Abstract: Embodiments of the present disclosure relate to a method and apparatus for generating a dialogue model. The method may include: acquiring a corpus sample set, a corpus sample including input information and target response information; classifying corpus samples in the corpus sample set, setting discrete hidden variables for the corpus samples based on a classification result to generate a training sample set, a training sample including the input information, the target response information, and a discrete hidden variable; and training a preset neural network using the training sample set to obtain the dialogue model, the dialogue model being used to represent a corresponding relationship between inputted input information and outputted target response information.
-
6.
公开(公告)号:US20210192284A1
公开(公告)日:2021-06-24
申请号:US16901940
申请日:2020-06-15
Inventor: Hao XIONG , Zhongjun HE , Zhi LI , Hua WU , Haifeng WANG
IPC: G06K9/62 , G06F40/117
Abstract: The present disclosure provides an end-to-end model training method and apparatus, which relates to a field of artificial intelligence technologies. The method includes: obtaining training data containing a plurality of training samples, in which the plurality of training samples include an original sequence, a target sequence and a corresponding tag list, the tag list includes importance tags in the target sequence and avoidance tags corresponding to the importance tags, and the avoidance tags are irrelevant tags corresponding to the importance tags; and adopting the training data to train a preset end-to-end model until a value of a preset optimization target function is smaller than a preset threshold.
-
公开(公告)号:US20210182498A1
公开(公告)日:2021-06-17
申请号:US16885358
申请日:2020-05-28
Inventor: Yu SUN , Haifeng WANG , Shuohuan WANG , Yukun LI , Shikun FENG , Hao TIAN , Hua WU
Abstract: The present disclosure provides a method, apparatus, electronic device and storage medium for processing a semantic representation model, and relates to the field of artificial intelligence technologies. A specific implementation solution is: collecting a training corpus set including a plurality of training corpuses; training the semantic representation model using the training corpus set based on at least one of lexicon, grammar and semantics. In the present disclosure, by building the unsupervised or weakly-supervised training task at three different levels, namely, lexicon, grammar and semantics, the semantic representation model is enabled to learn knowledge at levels of lexicon, grammar and semantics from massive data, enhance the capability of universal semantic representation and improve the processing effect of the NLP task.
-
8.
公开(公告)号:US20220171941A1
公开(公告)日:2022-06-02
申请号:US17348104
申请日:2021-06-15
Inventor: Xuan OUYANG , Shuohuan WANG , Chao PANG , Yu SUN , Hao TIAN , Hua WU , Haifeng WANG
Abstract: The present disclosure provides a multi-lingual model training method, apparatus, electronic device and readable storage medium and relates to the technical field of deep learning and natural language processing. A technical solution of the present disclosure when training the multi-lingual model is: obtaining training corpuses comprising a plurality of bilingual corpuses and a plurality of monolingual corpuses; training a multi-lingual model with a first training task by using the plurality of bilingual corpuses; training the multi-lingual model with a second training task by using the plurality of monolingual corpuses; and completing the training of the multi-lingual model in a case of determining that loss functions of the first training task and second training task converge. In the present disclosure, the multi-lingual model can be enabled to achieve semantic interaction between different languages and improve the accuracy of the multi-lingual model in learning the semantic representations of the multi-lingual model.
-
公开(公告)号:US20210406480A1
公开(公告)日:2021-12-30
申请号:US17475073
申请日:2021-09-14
Inventor: Fan WANG , Siqi BAO , Xinxian HUANG , Hua WU , Jingzhou HE
IPC: G06F40/40 , G06F16/33 , G06F16/332 , G06N7/00
Abstract: The disclosure discloses a method for generating a conversation, an electronic device, and a storage medium. The detailed implementation includes: obtaining a current conversation and historical conversations of the current conversation; selecting multiple reference historical conversations from the historical conversations and adding the multiple reference historical conversations to a temporary conversation set; and generating reply information of the current conversation based on the current conversation and the temporary conversation set.
-
10.
公开(公告)号:US20210294975A1
公开(公告)日:2021-09-23
申请号:US17015411
申请日:2020-09-09
Inventor: Xinchao XU , Haifeng WANG , Hua WU , Zhanyi LIU
IPC: G06F40/284 , G10L15/06 , G06N3/08 , G10L15/04 , G06F40/274
Abstract: A method, an electronic device and a readable storage medium for creating a label marking model are disclosed. According to an embodiment, the method for creating the label marking model includes: obtaining text data and determining a word or phrase to be marked in the text data; according to the word or phrase to be marked, constructing a first training sample of the text data corresponding to a word or phrase replacing task and a second training sample corresponding to a label marking task; training a neural network model with a plurality of the first training samples and a plurality of the second training samples, respectively, until a loss function of the word or phrase replacing task and a loss function of the label marking task satisfy a preset condition, to obtain the label marking model. The technical solution may improve the accuracy of the label marking model.
-
-
-
-
-
-
-
-
-