Invention Publication
- Patent Title: METHOD AND APPARATUS FOR PRE-TRAINING SEMANTIC REPRESENTATION MODEL AND ELECTRONIC DEVICE
-
Application No.: US18051594Application Date: 2022-11-01
-
Publication No.: US20230147550A1Publication Date: 2023-05-11
- Inventor: Dongliang HE , Errui DING
- Applicant: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
- Applicant Address: CN Beijing
- Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
- Current Assignee: BEIJING BAIDU NETCOM SCIENCE TECHNOLOGY CO., LTD.
- Current Assignee Address: CN Beijing
- Priority: CN 2111307885.6 2021.11.05
- Main IPC: G06V10/774
- IPC: G06V10/774 ; G06V20/40 ; G06F40/30 ; G06V30/19

Abstract:
A method for pre-training a semantic representation model includes: for each video-text pair in pre-training data, determining a mask image sequence, a mask character sequence, and a mask image-character sequence of the video-text pair; determining a plurality of feature sequences and mask position prediction results respectively corresponding to the plurality of feature sequences by inputting the mask image sequence, the mask character sequence, and the mask image-character sequence into an initial semantic representation model; and building a loss function based on the plurality of feature sequences, the mask position prediction results respectively corresponding to the plurality of feature sequences and true mask position results, and adjusting coefficients of the semantic representation model to realize training.
Information query