-
公开(公告)号:US20220179485A1
公开(公告)日:2022-06-09
申请号:US17545932
申请日:2021-12-08
Applicant: BEIHANG UNIVERSITY
Inventor: FENG LU , YIWEI BAO , QINPING ZHAO
Abstract: The present application provides a gaze point estimation method, device, and an electronic device. The method includes: acquiring user image data; acquiring a facial feature vector according to a preset first convolutional neural network and the facial image; acquiring a position feature vector according to a preset first fully connected network and the position data; acquiring a binocular fusion feature vector according to a preset eye feature fusion network, the left-eye image and the right-eye image; and acquiring position information about a gaze point of a user according to a preset second fully connected network, the facial feature vector, the position feature vector, and the binocular fusion feature vector. In this technical solution, relation between eye images and face images is utilized to achieve accurate gaze point estimation.
-
2.
公开(公告)号:US20190080455A1
公开(公告)日:2019-03-14
申请号:US15871878
申请日:2018-01-15
Applicant: BEIHANG UNIVERSITY
Inventor: XIAOWU CHEN , JIA LI , YAFEI SONG , YIFAN ZHAO , QINPING ZHAO
Abstract: Embodiments of the present invention provide a method and a device for three-dimensional feature-embedded image object component-level semantic segmentation, the method includes: acquiring three-dimensional feature information of a target two-dimensional image; performing a component-level semantic segmentation on the target two-dimensional image according to the three-dimensional feature information of the target two-dimensional image and two-dimensional feature information of the target two-dimensional image. In the technical solution of the present application, not only the two-dimensional feature information of the image but also the three-dimensional feature information of the image are taken into consideration when performing the component-level semantic segmentation on the image, thereby improving the accuracy of the image component-level semantic segmentation.
-
3.
公开(公告)号:US20180247126A1
公开(公告)日:2018-08-30
申请号:US15719480
申请日:2017-09-28
Applicant: BEIHANG UNIVERSITY
Inventor: JIA LI , XIAOWU CHEN , BIN ZHOU , QINPING ZHAO , CHANGQUN XIA , ANLIN ZHENG , YU ZHANG
CPC classification number: G06K9/00718 , G06K9/00744 , G06K9/00765 , G06K9/4638 , G06K9/4642 , G06K9/6207 , G06K9/6215 , G06K9/6259 , G06K9/6265 , G06K9/6267
Abstract: Provided is a method and a system for detecting and segmenting primary video objects with neighborhood reversibility, including: dividing each video frame of a video into super pixel blocks; representing each super pixel block with visual features; constructing and training a deep neural network to predict the initial foreground value for each super pixel block in the spatial domain; constructing a neighborhood reversible matrix and transmitting the initial foreground value, constructing an iterative optimization problem and resolving the final foreground value in the temporal spatial domain; performing pixel level transformation on the final foreground value; optimizing the final foreground value for the pixel using morphological smoothing operations; determining whether the pixel belongs to the primary video objects according to the final foreground value. The present disclosure does not require to set a priori assumption for processing a video, and is especially suitable for big data sets including complicated scenarios.
-
4.
公开(公告)号:US20180204378A1
公开(公告)日:2018-07-19
申请号:US15835408
申请日:2017-12-07
Applicant: BEIHANG UNIVERSITY
Inventor: JUN PAN , SHIZENG YAN , QINPING ZHAO
Abstract: Disclosed is a method for real-time cutting of digital organ based on a metaball model and a hybrid driving method, including a cutting procedure for driving a model using a position-based dynamics and a meshless method, a cutting mode which begins from a metaball driven by the position-based dynamics, proceeds to a point set driven by the meshless method and then create a new metaball. The method includes: a preprocessing procedure which performs an initialization operation while reading a model file; a deforming procedure which drives a model using a method based on the position-based dynamics; a cutting procedure which drives the model using the hybrid driving method and performs cutting using said cutting mode; and a rendering procedure which renders the model during the second and third procedures.
-
公开(公告)号:US20200349750A1
公开(公告)日:2020-11-05
申请号:US16725634
申请日:2019-12-23
Applicant: BEIHANG UNIVERSITY
Inventor: FENG LU , QINPING ZHAO
Abstract: Embodiments of the present disclosure provide a method and an apparatus for eye movement synthesis, the method including: obtaining eye movement feature data and speech feature data, wherein the eye movement feature data reflects an eye movement behavior, and the speech feature data reflects a voice feature; obtaining a driving model according to the eye movement feature data and the speech feature data, wherein the driving model is configured to indicate an association between the eye movement feature data and the speech feature data; synthesizing an eye movement of a virtual human according to speech input data and the driving model and controlling the virtual human to exhibit the synthesized eye movement. The embodiment makes the virtual human to exhibit an eye movement corresponding to the voice data according to the eye movement feature data and the speech feature data, thereby improving the authenticity in the interaction.
-
公开(公告)号:US20180211393A1
公开(公告)日:2018-07-26
申请号:US15710791
申请日:2017-09-20
Applicant: BEIHANG UNIVERSITY
Inventor: XIAOWU CHEN , YU ZHANG , JIA LI , WEI TENG , HAOKUN SONG , QINPING ZHAO
CPC classification number: G06T7/194 , G06K9/00744 , G06K9/34 , G06K9/38 , G06K9/4604 , G06K9/6202 , G06T7/11 , G06T7/143 , G06T7/174 , G06T2207/10016
Abstract: The disclosure involves an image guided video semantic object segmentation method and apparatus, locate a target object in a sample image to obtain an object sample; extract a candidate region from each frame; match multiple candidate regions extracted from the each frame with the object sample to obtain a similarity rating of each candidate region; rank the similarity rating of each candidate region to select a predefined candidate region number of high rating candidate region ranked by the similarity; preliminarily segment a foreground and a background from the selected high rating candidate region; construct an optimization function for the preliminarily segmented foreground and background; solve the optimization function to obtain a optimal candidate region set; and propagate a preliminary foreground segmentation corresponding to the optimal candidate region to an entire video to obtain a semantic object segmentation of the input video.
-
7.
公开(公告)号:US20180204088A1
公开(公告)日:2018-07-19
申请号:US15792712
申请日:2017-10-24
Applicant: BEIHANG UNIVERSITY
Inventor: XIAOWU CHEN , CHANGQUN XIA , JIA LI , QINPING ZHAO
CPC classification number: G06K9/4671 , G06K9/4638 , G06K9/6262 , G06T7/11 , G06T7/143 , G06T7/194 , G06T7/70 , G06T7/80 , G06T2207/10024 , G06T2207/20016 , G06T2207/20076 , G06T2207/20156
Abstract: Provided is a method for salient object segmentation of an image by aggregating a multi-linear exemplar regressors, including: analyzing and summarizing visual attributes and features of a salient object and a non-salient object using background prior and constructing a quadratic optimization problem, calculating an initial saliency probability map, selecting a most trusted foreground and a background seed point, performing manifold preserving foreground propagation, generating a final foreground probability map; generating a candidate object set for the image via an objectness adopting proposal, using a shape feature, a foregroundness and an attention feature to characterize each candidate object, training the linear exemplar regressors for each training image to characterize a particular saliency pattern of the image; aggregating a plurality of linear exemplar regressors, calculating saliency values for the candidate object set of a test image, and forming an image salient object segmentation model capable of processing various complex scenarios.
-
8.
公开(公告)号:US20180018539A1
公开(公告)日:2018-01-18
申请号:US15448546
申请日:2017-03-02
Applicant: BEIHANG UNIVERSITY
Inventor: XIAOWU CHEN , YAFEI SONG , JIA LI , QINPING ZHAO , XIAOGANG WANG
CPC classification number: G06T5/001 , G06K9/00973 , G06K9/4628 , G06K9/623 , G06K9/627 , G06N3/04 , G06N3/0454 , G06N3/08 , G06N3/084 , G06N5/003 , G06N20/20 , G06T1/20 , G06T2200/28 , G06T2207/10024 , G06T2207/20081 , G06T2207/20084 , G06T2207/30192
Abstract: Present invention provides a ranking convolutional neural network constructing method and an image processing method and apparatus thereof. The ranking convolutional neural network includes a ranking layer that is configured to sort an output of a previous layer of the ranking layer, generate an output of the ranking layer according to the sorted output, and output the output of the ranking layer to a next layer of the ranking layer. Using the ranking convolutional neural network enables obtaining an output feature corresponding to the input feature image through automatic learning. Compared with prior art methods that obtain features through manual calculation, the method of the present invention is superior in terms of reflecting the objective laws contained by the patterns of the actual scene. When applied to the field of image processing, the method can significantly improve the effect of image processing.
-
公开(公告)号:US20220382553A1
公开(公告)日:2022-12-01
申请号:US17546993
申请日:2021-12-09
Applicant: BEIHANG UNIVERSITY
Inventor: JIA LI , YIFAN ZHAO , DINGFENG SHI , QINPING ZHAO
Abstract: Embodiments of the present disclosure provides a fine-grained image recognition method and apparatus using graph structure represented high-order relation discovery, wherein the method includes: inputting an image to be classified into a convolutional neural network feature extractor with multiple stages, extracting two layers of network feature graphs in the last stage, constructing a hybrid high-order attention module according to the network feature graphs, and forming a high-order feature vector pool according to the hybrid high-order attention module, using each vector in the vector pool as a node, and utilizing semantic similarity among high-order features to form representative vector nodes in groups, and performing global pooling on the representative vector nodes to obtain classification vectors, and obtaining a fine-grained classification result through a fully connected layer and a classifier based on the classification vectors.
-
公开(公告)号:US20190087964A1
公开(公告)日:2019-03-21
申请号:US16132345
申请日:2018-09-14
Applicant: BEIHANG UNIVERSITY
Inventor: XIAOWU CHEN , XIAOGANG WANG , BIN ZHOU , HAIYUE FANG , QINPING ZHAO
Abstract: The present application provides a method and an apparatus for parsing and processing a three-dimensional CAD model, where the method includes: determining three kinds of adjacency relation information for each component in the three-dimensional model; performing aggregation processing on all components of the three-dimensional model, and generating three part hypothesis sets for the three-dimensional model; performing voxelization expression processing on each part hypothesis in each part hypothesis set, and generating voxelization information for each part hypothesis; inputting voxelization information of all part hypotheses in each part hypothesis set into an identification model to obtain a confidence score and a semantic category probability distribution for each part hypothesis; and constructing, according to the confidence score and the semantic category probability distribution of each part hypothesis in the part hypothesis sets, a high-order conditional random field model, and obtaining a semantic category analysis result for each component in the three-dimensional model.
-
-
-
-
-
-
-
-
-