-
公开(公告)号:US20240177318A1
公开(公告)日:2024-05-30
申请号:US18107173
申请日:2023-02-08
Applicant: Samsung Electronics Co., Ltd.
Inventor: Mostafa EL-KHAMY , Hai SU , Qingfeng LIU
IPC: G06T7/12
CPC classification number: G06T7/12 , G06T2207/10016 , G06T2207/20081
Abstract: Disclosed is a method including receiving, in a semantic segmentation network, input data from a plurality of frames, computing a ground truth label on the plurality of frames, generating a ground truth temporal semantic boundary map from the ground truth label on the plurality of frames, generating a predicted temporal semantic boundary map based on an output of the input data, and determining a loss based on the ground truth temporal semantic boundary map and the predicted temporal semantic boundary map.
-
公开(公告)号:US20250069247A1
公开(公告)日:2025-02-27
申请号:US18943520
申请日:2024-11-11
Applicant: SAMSUNG ELECTRONICS CO., LTD.
Inventor: Haoyu REN , Mostafa EL-KHAMY , Jungwon LEE , Hai SU , Qingfeng LIU
Abstract: Methods and systems for performing video prediction, including obtaining an input frame from among a plurality of frames included in an input video; extracting a first feature map by providing the input frame to a first plurality of feature extraction layers and a first strided convolutional layer included in an encoder; providing the first feature map and at least one neighboring first feature map corresponding to at least one neighboring frame to a first fusion module included in the encoder; fusing the first feature map with the at least one neighboring first feature map to generate a fused first feature map using the first fusion module; generating a prediction corresponding to the input frame based on the fused first feature map using a decoder; and performing a video prediction task using the prediction.
-
公开(公告)号:US20220301128A1
公开(公告)日:2022-09-22
申请号:US17563012
申请日:2021-12-27
Applicant: Samsung Electronics Co., Ltd.
Inventor: Qingfeng LIU , Hai SU , Mostafa EL-KHAMY
Abstract: A method of image processing includes: determining a first feature, wherein the first feature has a dimensionality D1; determining a second feature, wherein the second feature has a dimensionality D2 and is based on an output of a feature extraction network; generating a third feature by processing the first feature, the third feature having a dimensionality D3; generating a guidance by processing the second feature, the guidance having the dimensionality D3; generating a filter output by applying a deep guided filter (DGF) to the third feature using the guidance; generating a map based on the filter output; and outputting a processed image based on the map.
-
-