-
公开(公告)号:US20240104894A1
公开(公告)日:2024-03-28
申请号:US17949078
申请日:2022-09-20
Applicant: Lemon Inc.
IPC: G06V10/764 , G06V10/72 , G06V10/771
CPC classification number: G06V10/764 , G06V10/72 , G06V10/771
Abstract: A method is proposed for sample processing. A first group of data are received, here data in the first group of data comprises a sample and a classification of the sample, and the classification belonging to a first group of classifications in a plurality of classifications associated with the data. A plurality of data with the classification are selected from the first group of data. A first and a second loss function are determined for training a classification model that represents an association relationship between samples and classifications of the samples based on a plurality of samples comprised in the plurality of data and the classification, the first and second loss functions represent classification accuracy and a feature distribution for the classification model. The classification model is trained based on the first and second loss functions. Therefore, the accuracy of the classification model may be increased.
-
公开(公告)号:US20240144656A1
公开(公告)日:2024-05-02
申请号:US18394249
申请日:2023-12-22
Applicant: Lemon Inc. , Beijing Youzhuju Network Technology Co., Ltd.
Inventor: Song Bai , Junhao Zhang , Heng Wang , Rui Yan , Chuhui Xue , Wenqing Zhang
IPC: G06V10/774 , G06V10/40 , G06V10/74 , G06V10/772 , G06V10/82
CPC classification number: G06V10/774 , G06V10/40 , G06V10/761 , G06V10/772 , G06V10/82
Abstract: A method, apparatus, device, and medium for image processing is provided. The method includes generating, using an image generation process, a first set of synthetic images based on a first set of codes associated with the first image class in a codebook and based on a first class feature associated with a first image class; generating, using a feature extraction process, a first set of reference features based on the first set of synthetic images and generating a first set of target features based on a plurality of sets of training images belonging to the first image class in a training image set; and updating the image generation process and the codebook according to at least a first training objective to reduce a difference between each reference feature in the first set of reference features and a corresponding target feature in the first set of target features.
-
3.
公开(公告)号:US20240265571A1
公开(公告)日:2024-08-08
申请号:US18432448
申请日:2024-02-05
Applicant: Lemon Inc. , Beijing Youzhuju Network Technology Co., Ltd.
Inventor: Song Bai , Qihao Liu , Junfeng Wu , Yi Jiang
CPC classification number: G06T7/74 , G06T7/20 , G06T2207/10016
Abstract: A method, device, and medium for tracking a target object in a video based on an instance motion are provided. In one method, for a set of previous frames prior to a target frame in the video, a set of previous positions of the target object in the set of previous frames is obtained respectively. Based on the set of previous positions, a predicted value of a position of the target object in the target frame is determined with a motion model. A measured value of a position of an object in the target frame is determined. Based on a similarity between the predicted value and the measured value, the target object is tracked in the video.
-
公开(公告)号:US20240144664A1
公开(公告)日:2024-05-02
申请号:US18393238
申请日:2023-12-21
Applicant: Lemon Inc. , Beijing Youzhuju Network Technology Co., Ltd.
Inventor: Song Bai , Rui Yan , Heng Wang , Junhao Zhang , Chuhui Xue , Wenqing Zhang
CPC classification number: G06V10/82 , G06V10/467
Abstract: Embodiments of the present disclosure provide a solution for multimodal data processing. A method comprises: obtaining image data and text data; and extracting a target visual feature of image data and a target textual feature of text data using a feature extraction model. The feature extraction model comprises alternatively deployed cross-modal encoding parts and visual encoding parts. The extracting comprises: performing, using a first cross-modal encoding part of the feature extraction model, cross-modal feature encoding on a first intermediate visual feature of the image data and a first intermediate textual feature of the text data, to obtain a second intermediate visual feature and a second intermediate textual feature; performing, using a first visual encoding part of the feature extraction model, visual modal feature encoding on the second intermediate visual feature, to obtain a third intermediate visual feature.
-
公开(公告)号:US12254707B2
公开(公告)日:2025-03-18
申请号:US17955285
申请日:2022-09-28
Applicant: Lemon Inc. , Beijing Youzhuju Network Technology Co., Ltd.
Inventor: Chuhui Xue , Wenqing Zhang , Yu Hao , Song Bai
Abstract: Embodiments of the present disclosure relate to a method, device and computer readable storage medium of scene text detection. In the method, a first visual representation of a first image is generated with an image encoding process. A first textual representation of a first text unit in the first image is generated with a text encoding process based on a first plurality of symbols obtained by masking a first symbol of a plurality of symbols in the first text unit. A first prediction of the masked first symbol is determined with a decoding process based on the first visual and textual representations. At least the image encoding process is updating according to at least a first training objective to increase at least similarity of the first prediction and the masked first symbol.
-
公开(公告)号:US20240193412A1
公开(公告)日:2024-06-13
申请号:US18063843
申请日:2022-12-09
Applicant: Lemon Inc.
Inventor: Song Bai , Zhongcong Xu , Jiashi Feng , Jun Hao Liew , Wenqing Zhang
IPC: G06N3/08
CPC classification number: G06N3/08 , G06T2207/20081
Abstract: Generating a multi-dimensional video using a multi-dimensional video generative model for, including, but not limited to, at least one of static portrait animation, video reconstruction, or motion editing. The method including providing data into the multi-dimensionally aware generator of the multi-dimensional video generative model, and generating the multi-dimensional video from the data by the multi-dimensionally aware generator. The generating of the multi-dimensional video includes inverting the data into a latent space of the multi-dimensionally aware generator, synthesizing content of the multi-dimensional video using an appearance component of the multi-dimensionally aware generator and corresponding camera pose and formulating an intermediate appearance code, developing a synthesis layer for encoding a motion component of the multi-dimensionally aware generator at a plurality of timesteps and formulating an intermediate motion code, introducing temporal dynamics into the intermediate appearance code and the intermediate motion code, and generating multi-dimensionally aware spatio-temporal representations of the data.
-
公开(公告)号:US20220270353A1
公开(公告)日:2022-08-25
申请号:US17740211
申请日:2022-05-09
Applicant: Lemon Inc.
Inventor: Song Bai , Jieneng Chen , Shuyang Sun , Ju He , Bin Lu
IPC: G06V10/77 , G06V10/774 , G06V10/48 , G06V10/26 , G06N20/00
Abstract: Implementations of the present disclosure relate to methods, devices, and computer program products for data augmentation. In the method, mixed data is generated from first data and second data, and the mixed data comprises a first portion from the first data and a second portion from the second data. An attention map is obtained for the mixed data based on distributions of the first and second portions in the mixed data, here the attention map describes contributions of the first and second data to the mixed data. A label is determined for the mixed data based on the attention map and a first label for the first data and a second label for the second data. With these implementations, the label is determined based on the contributions of the first and second images in an accurate and effective way, and thus has a value that is much closer to the ground true.
-
-
-
-
-
-