-
公开(公告)号:US20230077258A1
公开(公告)日:2023-03-09
申请号:US17398673
申请日:2021-08-10
Applicant: Nvidia Corporation
Inventor: Maying Shen , Pavlo Molchanov , Hongxu Yin , Lei Mao , Jianna Liu , Jose Manuel Alvarez Lopez
Abstract: Apparatuses, systems, and techniques are presented to simplify neural networks. In at least one embodiment, one or more portions of one or more neural networks are cause to be removed based, at least in part, on one or more performance metrics of the one or more neural networks.
-
公开(公告)号:US20240119291A1
公开(公告)日:2024-04-11
申请号:US18203552
申请日:2023-05-30
Applicant: NVIDIA Corporation
Inventor: Jose M. Alvarez Lopez , Pavlo Molchanov , Hongxu Yin , Maying Shen , Lei Mao , Xinglong Sun
IPC: G06N3/082 , G06N3/0495
CPC classification number: G06N3/082 , G06N3/0495
Abstract: Machine learning is a process that learns a neural network model from a given dataset, where the model can then be used to make a prediction about new data. In order to reduce the size, computation, and latency of a neural network model, a compression technique can be employed which includes model sparsification. To avoid the negative consequences of pruning a fully pretrained neural network model and on the other hand of training a sparse model in the first place without any recovery option, the present disclosure provides a dynamic neural network model sparsification process which allows for recovery of previously pruned parts to improve the quality of the sparse neural network model.
-
公开(公告)号:US20240253217A1
公开(公告)日:2024-08-01
申请号:US18538248
申请日:2023-12-13
Applicant: NVIDIA Corporation
Inventor: Arash Vahdat , Hongxu Yin , Jan Kautz , Jiaming Song , Ming-Yu Liu , Morteza Mardani , Qinsheng Zhang
IPC: B25J9/16
CPC classification number: B25J9/163 , B25J9/1664 , B25J9/1697
Abstract: Apparatuses, systems, and techniques to calculate a combined loss value based on applying one or more loss functions to the plurality of samples generated by a diffusion model to update the samples to determine a synthesized motions of one or more objects.
-
公开(公告)号:US20220292360A1
公开(公告)日:2022-09-15
申请号:US17201768
申请日:2021-03-15
Applicant: NVIDIA Corporation
Inventor: Maying Shen , Pavlo Molchanov , Hongxu Yin , Jose Manuel Alvarez Lopez
Abstract: Apparatuses, systems, and techniques to remove one or more nodes of a neural network. In at least one embodiment, one or more nodes of a neural network are removed, based on, for example, whether the one or more nodes are likely to affect performance of the neural network.
-
公开(公告)号:US20240127067A1
公开(公告)日:2024-04-18
申请号:US18459083
申请日:2023-08-31
Applicant: NVIDIA Corporation
Inventor: Annamarie Bair , Hongxu Yin , Pavlo Molchanov , Maying Shen , Jose Manuel Alvarez Lopez
IPC: G06N3/082
CPC classification number: G06N3/082
Abstract: Systems and methods are disclosed for improving natural robustness of sparse neural networks. Pruning a dense neural network may improve inference speed and reduces the memory footprint and energy consumption of the resulting sparse neural network while maintaining a desired level of accuracy. In real-world scenarios in which sparse neural networks deployed in autonomous vehicles perform tasks such as object detection and classification for acquired inputs (images), the neural networks need to be robust to new environments, weather conditions, camera effects, etc. Applying sharpness-aware minimization (SAM) optimization during training of the sparse neural network improves performance for out of distribution (OOD) images compared with using conventional stochastic gradient descent (SGD) optimization. SAM optimizes a neural network to find a flat minimum: a region that both has a small loss value, but that also lies within a region of low loss.
-
公开(公告)号:US20230394781A1
公开(公告)日:2023-12-07
申请号:US18083397
申请日:2022-12-16
Applicant: NVIDIA Corporation
Inventor: Ali Hatamizadeh , Hongxu Yin , Jan Kautz , Pavlo Molchanov
CPC classification number: G06V10/42 , G06V10/44 , G06V10/82 , G06T3/40 , G06V10/7715
Abstract: Vision transformers are deep learning models that employ a self-attention mechanism to obtain feature representations for an input image. To date, the configuration of vision transformers has limited the self-attention computation to a local window of the input image, such that short-range dependencies are modeled in the output. The present disclosure provides a vision transformer that captures global context, and that is therefore able to model long-range dependencies in its output.
-
公开(公告)号:US20230080247A1
公开(公告)日:2023-03-16
申请号:US17551005
申请日:2021-12-14
Applicant: NVIDIA Corporation
Inventor: Hongxu Yin , Huanrui Yang , Pavlo Molchanov , Jan Kautz
Abstract: A vision transformer is a deep learning model used to perform vision processing tasks such as image recognition. Vision transformers are currently designed with a plurality of same-size blocks that perform the vision processing tasks. However, some portions of these blocks are unnecessary and not only slow down the vision transformer but use more memory than required. In response, parameters of these blocks are analyzed to determine a score for each parameter, and if the score falls below a threshold, the parameter is removed from the associated block. This reduces a size of the resulting vision transformer, which reduces unnecessary memory usage and increases performance.
-
公开(公告)号:US20240185034A1
公开(公告)日:2024-06-06
申请号:US18130648
申请日:2023-04-04
Applicant: NVIDIA Corporation
Inventor: Ali Hatamizadeh , Gregory Heinrich , Hongxu Yin , Jose Manuel Alvarez Lopez , Jan Kautz , Pavlo Molchanov
IPC: G06N3/0455 , G06N3/0464 , G06N3/08
CPC classification number: G06N3/0455 , G06N3/0464 , G06N3/08
Abstract: Apparatuses, systems, and techniques of using one or more machine learning processes (e.g., neural network(s)) to process data (e.g., using hierarchical self-attention). In at least one embodiment, image data is classified using hierarchical self-attention generated using carrier tokens that are associated with windowed subregions of the image data, and local attention generated using local tokens within the windowed subregions and the carrier tokens.
-
公开(公告)号:US20240096115A1
公开(公告)日:2024-03-21
申请号:US18243555
申请日:2023-09-07
Applicant: NVIDIA Corporation
Inventor: Pavlo Molchanov , Jan Kautz , Arash Vahdat , Hongxu Yin , Paul Micaelli
CPC classification number: G06V20/597 , G06T7/70 , G06V10/82 , G06V20/70 , G06V40/171 , G06T2207/30201 , G06V2201/07
Abstract: Landmark detection refers to the detection of landmarks within an image or a video, and is used in many computer vision tasks such emotion recognition, face identity verification, hand tracking, gesture recognition, and eye gaze tracking. Current landmark detection methods rely on a cascaded computation through cascaded networks or an ensemble of multiple models, which starts with an initial guess of the landmarks and iteratively produces corrected landmarks which match the input more finely. However, the iterations required by current methods typically increase the training memory cost linearly, and do not have an obvious stopping criteria. Moreover, these methods tend to exhibit jitter in landmark detection results for video. The present disclosure improves current landmark detection methods by providing landmark detection using an iterative neural network. Furthermore, when detecting landmarks in video, the present disclosure provides for a reduction in jitter due to reuse of previous hidden states from previous frames.
-
公开(公告)号:US20220284283A1
公开(公告)日:2022-09-08
申请号:US17195451
申请日:2021-03-08
Applicant: NVIDIA Corporation
Inventor: Hongxu Yin , Pavlo Molchanov , Jose Manuel Alvarez Lopez , Xin Dong
Abstract: Apparatuses, systems, and techniques to invert a neural network. In at least one embodiment, one or more neural network layers are inverted and, in at least one embodiment, loaded in reverse order.
-
-
-
-
-
-
-
-
-