-
1.
公开(公告)号:US20230206522A1
公开(公告)日:2023-06-29
申请号:US18111958
申请日:2023-02-21
Inventor: Licheng TANG , Jiaming LIU , Taizhang SHANG
CPC classification number: G06T11/203 , G06V30/19147 , G06V30/19127 , G06V30/22
Abstract: A training method for a handwritten text image generation model includes: obtaining training data including a sample content image, a first sample handwritten text image and a second sample handwritten text image, constructing an initial training model; obtaining a first predicted handwritten text image by inputting the sample content image and the second sample handwritten text image into an initial handwritten text image generation model of the initial training model; obtaining a second predicted handwritten text image by inputting the sample content image and the first sample handwritten text image into an initial handwritten text image reconstruction model of the initial training model; training the initial training model according to the first and second predicted handwritten text images and the first sample handwritten text image; and determining a handwritten text image generation model of the training model after training as a target handwritten text image generation model.
-
公开(公告)号:US20220180650A1
公开(公告)日:2022-06-09
申请号:US17683514
申请日:2022-03-01
Inventor: Licheng TANG , Jiaming LIU
IPC: G06V30/244 , G06V10/774 , G06V10/82 , G06V10/74
Abstract: A method of generating a font database, and a method of training a neural network model are provided, which relate to a field of artificial intelligence, in particular to a computer vision and deep learning technology. The method of generating the font database includes: determining, by using a trained similarity comparison model, a basic font database most similar to handwriting font data of a target user in a plurality of basic font databases as a candidate font database; and adjusting, by using a trained basic font database model for generating the candidate font database, the handwriting font data of the target user, so as to obtain a target font database for the target user.
-
公开(公告)号:US20220148239A1
公开(公告)日:2022-05-12
申请号:US17583259
申请日:2022-01-25
Inventor: Jiaming LIU , Licheng TANG
IPC: G06T11/20
Abstract: A method for training a font generation model is described below. A source domain sample character and a target domain association character are input into a font generation network to obtain a target domain generation character. The target domain generation character and at least one of a target domain sample character or the target domain association character are input into a loss analysis network to obtain a first loss, and a parameter of the font generation model is adjusted according to the first loss. The source domain sample character and a random vector are input into the font generation network to obtain a random domain generation character. The random domain generation character and a random domain sample character are input into the loss analysis network to obtain a second loss, and the parameter is readjusted according to the second loss.
-
4.
公开(公告)号:US20220189083A1
公开(公告)日:2022-06-16
申请号:US17682232
申请日:2022-02-28
Inventor: Licheng TANG , Jiaming LIU
IPC: G06T11/20 , G06V30/19 , G06V30/244 , G06N20/00
Abstract: Provided is a training method for a character generation model, and a character generation method, apparatus and device, which relates to the technical field of artificial intelligences, particularly, the technical field of computer vision and deep learning. The specific implementation schemes are: a source domain sample word and a target domain style word are input into the character generation model to obtain a target domain generation word; the target domain generation word and a target domain sample word are input into a pre-trained character classification model to calculate a feature loss of the character generation model; and a parameter of the character generation model is adjusted according to the feature loss.
-
公开(公告)号:US20220147695A1
公开(公告)日:2022-05-12
申请号:US17583263
申请日:2022-01-25
Inventor: Jiaming LIU , Licheng TANG
IPC: G06F40/109 , G06V30/244 , G06V30/19
Abstract: A method for training a font generation model is described below. A source domain sample character and a target domain association character are input into an encoder of the font generation model to obtain a sample character content feature and an association character style feature. The sample character content feature and the association character style feature are input into an attention mechanism network to obtain a target domain style feature. The sample character content feature and the target domain style feature are input into a decoder to obtain a target domain generation character. The target domain generation character and at least one of a target domain sample character or the target domain association character are input into a loss analysis network of the font generation model to obtain a model loss, and a parameter of the font generation model is adjusted according to the model loss.
-
6.
公开(公告)号:US20220237935A1
公开(公告)日:2022-07-28
申请号:US17682099
申请日:2022-02-28
Inventor: Jiaming LIU , Licheng TANG
IPC: G06V30/19 , G06T11/20 , G06F40/109 , G06T11/60
Abstract: Provided are a method for training a font generation model, a method for establishing a font library, and a device. The method for training a font generation model includes the following steps: a source-domain sample character is input into the font generation model to obtain a first target-domain generated character; the first target-domain generated character and a preset target-domain sample character are input into a character classification model to obtain a first feature loss of the font generation model; the first target-domain generated character and the target-domain sample character are input into a font classification model to obtain a second feature loss of the font generation model; a target feature loss is determined according to the first feature loss and/or the second feature loss; and the model parameter of the font generation model is updated according to the target feature loss.
-
公开(公告)号:US20220188637A1
公开(公告)日:2022-06-16
申请号:US17683512
申请日:2022-03-01
Inventor: Jiaming LIU , Licheng TANG , Zhibin HONG
IPC: G06N3/08 , G06N3/04 , G06F40/109
Abstract: There are provided a method for training an adversarial network model, a method for building a character library, an electronic device and a storage medium, which relate to a field of artificial intelligence technology, in particular to a field of computer vision and deep learning technologies. The method includes: generating a generated character based on a content character sample having a base font and a style character sample having a style font and generating a reconstructed character based on the content character sample, by using a generation model; calculating a basic loss of the generation model based on the generated character and the reconstructed character, by using a discrimination model; calculating a character loss of the generation model through classifying the generated character by using a trained character classification model; and adjusting a parameter of the generation model based on the basic loss and the character loss.
-
公开(公告)号:US20240303774A1
公开(公告)日:2024-09-12
申请号:US18020918
申请日:2022-06-10
Inventor: Changyong SHU , Jiaming LIU , Zhibin HONG , Junyu HAN
CPC classification number: G06T5/50 , G06T5/60 , G06T7/40 , G06T7/55 , G06T7/90 , G06V40/172 , G06T2207/10024 , G06T2207/20081 , G06T2207/20084 , G06T2207/20221 , G06T2207/30201
Abstract: A method of processing an image, an electronic device and a storage medium. The method includes: generating a to-be-processed image according to a first target image and a second target image, where an identity information of an object in the to-be-processed image is matched with an identity information of an object in the first target image; generating a set of disentangled images according to the second target image and the to-be-processed image, where the set of disentangled images includes a head-disentangled image and a disentangled repair image; and generating a fusion image according to the set of disentangled images, where an identity information and a texture information of an object in the fusion image are matched with the identity information and the texture information of the object in the to-be-processed image, respectively, and a to-be-repaired information related to the object in the fusion image is repaired.
-
9.
公开(公告)号:US20230154077A1
公开(公告)日:2023-05-18
申请号:US17682295
申请日:2022-02-28
Inventor: Licheng TANG , Jiaming LIU
IPC: G06T11/20 , G06F40/109
CPC classification number: G06T11/203 , G06F40/109
Abstract: Provided is a training method for a character generation model. The training method for a character generation model includes: a first training sample is input into a target model to calculate a first loss, where the first training sample includes a first source domain sample word and a first target domain sample word, and content of the first source domain sample word is different from content of the first target domain sample word; a second training sample is input into the target model to calculate a second loss, where the second training sample includes a second source domain sample word and a second target domain sample word, content of the second source domain sample word is the same as content of the second target domain sample word; and a parameter of the character generation model is adjusted according to the first loss and the second loss.
-
10.
公开(公告)号:US20220189189A1
公开(公告)日:2022-06-16
申请号:US17683508
申请日:2022-03-01
Inventor: Licheng TANG , Jiaming LIU
IPC: G06V30/244 , G06V10/82 , G06V10/778 , G06V30/12 , G06V30/18 , G06V10/764 , G06N3/04
Abstract: A method of training a cycle generative networks model and a method of building a character library are provided, which relate to a field of artificial intelligence, in particular to a computer vision and deep learning technology, and which may be applied to a scene such as image processing and image recognition. A specific implementation scheme includes: inputting a source domain sample character into the cycle generative networks model to obtain a first target domain generated character; calculating a character error loss and a feature loss of the cycle generative networks model by inputting the first target domain generated character and a preset target domain sample character into a character classification model; and adjusting a parameter of the cycle generative networks model according to the character error loss and the feature loss. An electronic device and a storage medium are further provided.
-
-
-
-
-
-
-
-
-