-
公开(公告)号:US20240046072A1
公开(公告)日:2024-02-08
申请号:US18489730
申请日:2023-10-18
Applicant: Snap Inc.
Inventor: Linjie Yang , Jianchao Yang , Xuehan Xiong , Yanran Wang
Abstract: A modulated segmentation system can use a modulator network to emphasize spatial prior data of an object to track the object across multiple images. The modulated segmentation system can use a segmentation network that receives spatial prior data as intermediate data that improves segmentation accuracy. The segmentation network can further receive visual guide information from a visual guide network to increase tracking accuracy via segmentation.
-
公开(公告)号:US11743426B2
公开(公告)日:2023-08-29
申请号:US16992968
申请日:2020-08-13
Applicant: Snap Inc.
Inventor: Lidiia Bogdanovych , William Brendel , Samuel Edward Hare , Fedir Poliakov , Guohui Wang , Xuehan Xiong , Jianchao Yang , Linjie Yang
IPC: G06T7/194 , G06V10/82 , H04N7/14 , G06T7/11 , G06N3/08 , G06N3/04 , G06V30/242 , G06F18/214 , G06F18/24 , G06V30/19 , H04N5/445 , H04N5/76
CPC classification number: H04N7/147 , G06F18/214 , G06F18/24765 , G06N3/04 , G06N3/08 , G06T7/11 , G06T7/194 , G06V10/82 , G06V30/19173 , G06V30/242 , G06T2207/10016 , G06T2207/10024 , G06T2207/20024 , G06T2207/20081 , G06T2207/20084 , G06T2207/20221 , G06T2207/30201 , H04N5/44504 , H04N5/76 , H04N7/141
Abstract: A machine learning system can generate an image mask (e.g., a pixel mask) comprising pixel assignments for pixels. The pixels can he assigned to classes, including, for example, face, clothes, body skin, or hair. The machine learning system can be implemented. using a convolutional neural network that is configured to execute efficiently on computing devices having limited resources, such as mobile phones. The pixel mask can be used to more accurately display video effects interacting with a user or subject depicted in the image.
-
公开(公告)号:US10402689B1
公开(公告)日:2019-09-03
申请号:US15706057
申请日:2017-09-15
Applicant: Snap Inc.
Inventor: Lidiia Bogdanovych , William Brendel , Samuel Edward Hare , Fedir Poliakov , Guohui Wang , Xuehan Xiong , Jianchao Yang , Linjie Yang
IPC: G06T7/11 , G06K9/62 , G06K9/74 , G06T7/194 , G06N3/08 , G06N3/04 , G06K9/68 , H04N7/14 , H04N5/445 , H04N5/76
Abstract: A machine learning system can generate an image mask (e.g., a pixel mask) comprising pixel assignments for pixels. The pixels can be assigned to classes, including, for example, face, clothes, body skin, or hair. The machine learning system can be implemented using a convolutional neural network that is configured to execute efficiently on computing devices having limited resources, such as mobile phones. The pixel mask can be used to more accurately display video effects interacting with a user or subject depicted in the image.
-
公开(公告)号:US12075190B2
公开(公告)日:2024-08-27
申请号:US18221702
申请日:2023-07-13
Applicant: Snap Inc.
Inventor: Lidiia Bogdanovych , William Brendel , Samuel Edward Hare , Fedir Poliakov , Guohui Wang , Xuehan Xiong , Jianchao Yang , Linjie Yang
IPC: G06T7/11 , G06F18/214 , G06F18/24 , G06N3/04 , G06N3/08 , G06T7/194 , G06V10/82 , G06V30/19 , G06V30/242 , H04N7/14 , H04N5/445 , H04N5/76
CPC classification number: H04N7/147 , G06F18/214 , G06F18/24765 , G06N3/04 , G06N3/08 , G06T7/11 , G06T7/194 , G06V10/82 , G06V30/19173 , G06V30/242 , G06T2207/10016 , G06T2207/10024 , G06T2207/20024 , G06T2207/20081 , G06T2207/20084 , G06T2207/20221 , G06T2207/30201 , H04N5/44504 , H04N5/76 , H04N7/141
Abstract: A machine learning system can generate an image mask (e.g., a pixel mask) comprising pixel assignments for pixels. The pixels can be assigned to classes, including, for example, face, clothes, body skin, or hair. The machine learning system can be implemented using a convolutional neural network that is configured to execute efficiently on computing devices having limited resources, such as mobile phones. The pixel mask can be used to more accurately display video effects interacting with a user or subject depicted in the image.
-
公开(公告)号:US11727710B2
公开(公告)日:2023-08-15
申请号:US17508384
申请日:2021-10-22
Applicant: Snap Inc.
Inventor: Yuncheng Li , Linjie Yang , Ning Zhang , Zhengyuan Yang
CPC classification number: G06V40/107 , G06F17/15 , G06N3/04 , G06T7/136
Abstract: Segmentation of an image into individual body parts is performed based on a trained model. The model is trained with a plurality of training images, each training image representing a corresponding training figure. The model is also trained with a corresponding plurality of segmentations of the training figures. Each segmentation is generated by positioning body parts between defined positions of joints of the represented figure. The body parts are represented by body part templates obtained from a template library, with the templates defining characteristics of body parts represented by the templates.
-
公开(公告)号:US11551059B1
公开(公告)日:2023-01-10
申请号:US16192457
申请日:2018-11-15
Applicant: Snap Inc.
Inventor: Linjie Yang , Jianchao Yang , Xuehan Xiong , Yanran Wang
Abstract: A modulated segmentation system can use a modulator network to emphasize spatial prior data of an object to track the object across multiple images. The modulated segmentation system can use a segmentation network that receives spatial prior data as intermediate data that improves segmentation accuracy. The segmentation network can further receive visual guide information from a visual guide network to increase tracking accuracy via segmentation.
-
公开(公告)号:US11361489B2
公开(公告)日:2022-06-14
申请号:US16946346
申请日:2020-06-17
Applicant: Snap Inc.
Inventor: Linjie Yang , Kevin Dechau Tang , Jianchao Yang , Jia Li
Abstract: A dense captioning system and method is provided for analyzing an image to generate proposed bounding regions for a plurality of visual concepts within the image, generating a region feature for each proposed bounding region to generate a plurality of region features of the image, and determining a context feature for the image using a proposed bounding region that is a largest in size of the proposed bounding regions. For each region feature of the plurality of region features of the image, the dense captioning system and method further provides for analyzing the region feature to determine for the region feature a detection score that indicates a likelihood that the region feature comprises an actual object, and generating a caption for a visual concept in the image using the region feature and the context feature when a detection score is above a specified threshold value.
-
公开(公告)号:US10726306B1
公开(公告)日:2020-07-28
申请号:US16226035
申请日:2018-12-19
Applicant: Snap Inc.
Inventor: Linjie Yang , Kevin Dechau Tang , Jianchao Yang , Jia Li
Abstract: A dense captioning system and method is provided for analyzing an image to generate proposed bounding regions for a plurality of visual concepts within the image, generating a region feature for each proposed bounding region to generate a plurality of region features of the image, and determining a context feature for the image using a proposed bounding region that is a largest in size of the proposed bounding regions. For each region feature of the plurality of region features of the image, the dense captioning system and method further provides for analyzing the region feature to determine for the region feature a detection score that indicates a likelihood that the region feature comprises an actual object, and generating a caption for a visual concept in the image using the region feature and the context feature when a detection score is above a specified threshold value.
-
公开(公告)号:US20200050866A1
公开(公告)日:2020-02-13
申请号:US16654898
申请日:2019-10-16
Applicant: Snap Inc.
Inventor: Samuel Edward Hare , Fedir Poliakov , Guohui Wang , Xuehan Xiong , Jianchao Yang , Linjie Yang , Shah Tanmay Anilkumar
Abstract: A mobile device can generate real-time complex visual image effects using asynchronous processing pipeline. A first pipeline applies a complex image process, such as a neural network, to keyframes of a live image sequence. A second pipeline generates flow maps that describe feature transformations in the image sequence. The flow maps can be used to process non-keyframes on the fly. The processed keyframes and non-keyframes can be used to display a complex visual effect on the mobile device in real-time or near real-time.
-
公开(公告)号:US20190087660A1
公开(公告)日:2019-03-21
申请号:US15706096
申请日:2017-09-15
Applicant: Snap Inc.
Inventor: Samuel Edward Hare , Fedir Poliakov , Guohui Wang , Xuehan Xiong , Jianchao Yang , Linjie Yang , Shah Tanmay Anilkumar
Abstract: A mobile device can generate real-time complex visual image effects using asynchronous processing pipeline. A first pipeline applies a complex image process, such as a neural network, to keyframes of a live image sequence. A second pipeline generates flow maps that describe feature transformations in the image sequence. The flow maps can be used to process non-keyframes on the fly. The processed keyframes and non-keyframes can be used to display a complex visual effect on the mobile device in real-time or near real-time.
-
-
-
-
-
-
-
-
-