Invention Application
- Patent Title: MULTI-LINGUAL MODEL TRAINING METHOD, APPARATUS, ELECTRONIC DEVICE AND READABLE STORAGE MEDIUM
-
Application No.: US17348104Application Date: 2021-06-15
-
Publication No.: US20220171941A1Publication Date: 2022-06-02
- Inventor: Xuan OUYANG , Shuohuan WANG , Chao PANG , Yu SUN , Hao TIAN , Hua WU , Haifeng WANG
- Applicant: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
- Applicant Address: CN Beijing
- Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
- Current Assignee: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY CO., LTD.
- Current Assignee Address: CN Beijing
- Priority: CN202011387501.1 20201201
- Main IPC: G06F40/30
- IPC: G06F40/30 ; G06F40/58 ; G06N20/00

Abstract:
The present disclosure provides a multi-lingual model training method, apparatus, electronic device and readable storage medium and relates to the technical field of deep learning and natural language processing. A technical solution of the present disclosure when training the multi-lingual model is: obtaining training corpuses comprising a plurality of bilingual corpuses and a plurality of monolingual corpuses; training a multi-lingual model with a first training task by using the plurality of bilingual corpuses; training the multi-lingual model with a second training task by using the plurality of monolingual corpuses; and completing the training of the multi-lingual model in a case of determining that loss functions of the first training task and second training task converge. In the present disclosure, the multi-lingual model can be enabled to achieve semantic interaction between different languages and improve the accuracy of the multi-lingual model in learning the semantic representations of the multi-lingual model.
Public/Granted literature
- US11995405B2 Multi-lingual model training method, apparatus, electronic device and readable storage medium Public/Granted day:2024-05-28
Information query