-
公开(公告)号:US20220063605A1
公开(公告)日:2022-03-03
申请号:US17408911
申请日:2021-08-23
Applicant: NEC Laboratories America, Inc.
Inventor: Pan Ji , Buyu Liu , Bingbing Zhuang , Manmohan Chandraker , Xiangyu Chen
Abstract: A method provided for 3D object localization predicts pairs of 2D bounding boxes. Each pair corresponds to a detected object in each of the two consecutive input monocular images. The method generates, for each detected object, a relative motion estimation specifying a relative motion between the two images. The method constructs an object cost volume by aggregating temporal features from the two images using the pairs of 2D bounding boxes and the relative motion estimation to predict a range of object depth candidates and a confidence score for each object depth candidate and an object depth from the object depth candidates. The method updates the relative motion estimation based on the object cost volume and the object depth to provide a refined object motion and a refined object depth. The method reconstructs a 3D bounding box for each detected object based on the refined object motion and refined object depth.
-
公开(公告)号:US20210374468A1
公开(公告)日:2021-12-02
申请号:US17330832
申请日:2021-05-26
Applicant: NEC Laboratories America, Inc.
Inventor: Manmohan Chandraker , Ting Wang , Xiang Xu , Francesco Pittaluga , Gaurav Sharma , Yi-Hsuan Tsai , Masoud Faraki , Yuheng Chen , Yue Tian , Ming-Fang Huang , Jian Fang
Abstract: Methods and systems for training a neural network include generate an image of a mask. A copy of an image is generated from an original set of training data. The copy is altered to add the image of a mask to a face detected within the copy. An augmented set of training data is generated that includes the original set of training data and the altered copy. A neural network model is trained to recognize masked faces using the augmented set of training data.
-
公开(公告)号:US11055989B2
公开(公告)日:2021-07-06
申请号:US16051924
申请日:2018-08-01
Applicant: NEC Laboratories America, Inc.
Inventor: Kihyuk Sohn , Luan Tran , Xiang Yu , Manmohan Chandraker
IPC: G08G1/017 , G06K9/62 , G06K9/00 , G06N3/08 , G06N3/04 , G06N5/04 , G06T17/00 , G06N20/00 , G08G1/01 , G08G1/04 , G08G1/048 , G06K9/46
Abstract: Systems and methods for performing domain adaptation include collecting a labeled source image having a view of an object. Viewpoints of the object in the source image are synthesized to generate view augmented source images. Photometrics of each of the viewpoints of the object are adjusted to generate lighting and view augmented source images. Features are extracted from each of the lighting and view augmented source images with a first feature extractor and from captured images captured by an image capture device with a second feature extractor. The extracted features are classified using domain adaptation with domain adversarial learning between extracted features of the captured images and extracted features of the lighting and view augmented source images. Labeled target images are displayed corresponding to each of the captured images including labels corresponding to classifications of the extracted features of the captured images.
-
公开(公告)号:US20210150203A1
公开(公告)日:2021-05-20
申请号:US17096111
申请日:2020-11-12
Applicant: NEC Laboratories America, Inc.
Inventor: Buyu Liu , Bingbing Zhuang , Samuel Schulter , Manmohan Chandraker
Abstract: Systems and methods are provided for producing a road layout model. The method includes capturing digital images having a perspective view, converting each of the digital images into top-down images, and conveying a top-down image of time t to a neural network that performs a feature transform to form a feature map of time t. The method also includes transferring the feature map of the top-down image of time t to a feature transform module to warp the feature map to a time t+1, and conveying a top-down image of time t+1 to form a feature map of time t+1. The method also includes combining the warped feature map of time t with the feature map of time t+1 to form a combined feature map, transferring the combined feature map to a long short-term memory (LSTM) module to generate the road layout model, and displaying the road layout model.
-
公开(公告)号:US10991145B2
公开(公告)日:2021-04-27
申请号:US16673256
申请日:2019-11-04
Applicant: NEC Laboratories America, Inc.
Inventor: Xiang Yu , Feng-Ju Chang , Manmohan Chandraker
Abstract: A system is provided for pose-variant 3D facial attribute generation. A first stage has a hardware processor based 3D regression network for directly generating a space position map for a 3D shape and a camera perspective matrix from a single input image of a face and further having a rendering layer for rendering a partial texture map of the single input image based on the space position map and the camera perspective matrix. A second stage has a hardware processor based two-part stacked Generative Adversarial Network (GAN) including a Texture Completion GAN (TC-GAN) stacked with a 3D Attribute generation GAN (3DA-GAN). The TC-GAN completes the partial texture map to form a complete texture map based on the partial texture map and the space position map. The 3DA-GAN generates a target facial attribute for the single input image based on the complete texture map and the space position map.
-
公开(公告)号:US20210065391A1
公开(公告)日:2021-03-04
申请号:US16987705
申请日:2020-08-07
Applicant: NEC Laboratories America, Inc.
Inventor: Quoc-Huy Tran , Pan Ji , Manmohan Chandraker , Lokender Tiwari
Abstract: A method for improving geometry-based monocular structure from motion (SfM) by exploiting depth maps predicted by convolutional neural networks (CNNs) is presented. The method includes capturing a sequence of RGB images from an unlabeled monocular video stream obtained by a monocular camera, feeding the RGB images into a depth estimation/refinement module, outputting depth maps, feeding the depth maps and the RGB images to a pose estimation/refinement module, the depths maps and the RGB images collectively defining pseudo RGB-D images, outputting camera poses and point clouds, and constructing a 3D map of a surrounding environment displayed on a visualization device.
-
157.
公开(公告)号:US20200372614A1
公开(公告)日:2020-11-26
申请号:US16867805
申请日:2020-05-06
Applicant: NEC Laboratories America, Inc.
Inventor: Quoc-Huy Tran , Bingbing Zhuang , Pan Ji , Manmohan Chandraker
Abstract: A method for correcting blur effects is presented. The method includes generating a plurality of images from a camera, synthesizing blurred images from sharp image counterparts to generate training data to train a structure-and-motion-aware convolutional neural network (CNN), and predicting a camera motion and a depth map from a single blurred image by employing the structure-and-motion-aware CNN to remove blurring from the single blurred image.
-
公开(公告)号:US10832440B2
公开(公告)日:2020-11-10
申请号:US16115630
申请日:2018-08-29
Applicant: NEC Laboratories America, Inc.
Inventor: Samuel Schulter , Wongun Choi , Tuan Hung Vu , Manmohan Chandraker
Abstract: A computer-implemented method, system, and computer program product are provided for object detection utilizing an online flow guided memory network. The method includes receiving a plurality of videos, each of the plurality of videos including a plurality of frames. The method also includes generating, with a feature extraction network, a frame feature map for a current frame of the plurality of frames. The method additionally includes aggregating a memory feature map from the frame feature map and previous memory feature maps from previous frames on a plurality of time axes, with the plurality of time axes including a first time axis at a first frame increment and a second time axis at a second frame increment. The method further includes predicting, with a task network, an object from the memory feature map. The method also includes controlling an operation of a processor-based machine to react in accordance with the object.
-
159.
公开(公告)号:US20200286383A1
公开(公告)日:2020-09-10
申请号:US16787727
申请日:2020-02-11
Applicant: NEC Laboratories America, Inc.
Inventor: Quoc-Huy Tran , Samuel Schulter , Paul Vernaza , Buyu Liu , Pan Ji , Yi-Hsuan Tsai , Manmohan Chandraker
Abstract: A method is provided in an Advanced Driver-Assistance System (ADAS). The method extracts, from an input video stream including a plurality of images using a multi-task Convolutional Neural Network (CNN), shared features across different perception tasks. The perception tasks include object detection and other perception tasks. The method concurrently solves, using the multi-task CNN, the different perception tasks in a single pass by concurrently processing corresponding ones of the shared features by respective different branches of the multi-task CNN to provide a plurality of different perception task outputs. Each respective different branch corresponds to a respective one of the different perception tasks. The method forms a parametric representation of a driving scene as at least one top-view map responsive to the plurality of different perception task outputs. The method controls an operation of the vehicle for collision avoidance responsive to the at least one top-view map indicating an impending collision.
-
公开(公告)号:US10678256B2
公开(公告)日:2020-06-09
申请号:US16145621
申请日:2018-09-28
Applicant: NEC Laboratories America, Inc.
Inventor: Samuel Schulter , Paul Vernaza , Manmohan Chandraker , Menghua Zhai
IPC: G05D1/02 , G05D1/00 , G06T17/05 , G06T7/194 , G06T7/514 , G06T7/50 , G06K9/00 , G06T15/20 , G06T7/11 , G06T3/00 , G06T15/40
Abstract: Systems and methods for generating an occlusion-aware bird's eye view map of a road scene include identifying foreground objects and background objects in an input image to extract foreground features and background features corresponding to the foreground objects and the background objects, respectively. The foreground objects are masked from the input image with a mask. Occluded objects and depths of the occluded objects are inferred by predicting semantic features and depths in masked areas of the masked image according to contextual information related to the background features visible in the masked image. The foreground objects and the background objects are mapped to a three-dimensional space according to locations of each of the foreground objects, the background objects and occluded objects using the inferred depths. A bird's eye view is generated from the three-dimensional space and displayed with a display device.
-
-
-
-
-
-
-
-
-