-
公开(公告)号:US20220262105A1
公开(公告)日:2022-08-18
申请号:US17625313
申请日:2020-07-17
Applicant: Zongwei ZHOU , Vatsal SODHA , Md, Mahfuzur RAHMAN SIDDIQUEE , Ruibin FENG , Nima TAJBAKHSH , Jianming LIANG , Arizona Board of Regents on behalf of Arizona State University
Inventor: Zongwei Zhou , Vatsal Sodha , Md Mahfuzur Rahman Siddiquee , Ruibin Feng , Nima Tajbakhsh , Jianming Liang
IPC: G06V10/774 , G06V10/82 , G06V10/98 , G06V10/776
Abstract: Described herein are means for generating source models for transfer learning to application specific models used in the processing of medical imaging. In some embodiments, the method comprises: identifying a group of training samples, wherein each training sample in the group of training samples includes an image; for each training sample in the group of training samples: identifying an original patch of the image corresponding to the training sample; identifying one or more transformations to be applied to the original patch; generating a transformed patch by applying the one or more transformations to the identified patch; and training an encoder-decoder network using a group of transformed patches corresponding to the group of training samples, wherein the encoder-decoder network is trained to generate an approximation of the original patch from a corresponding transformed patch, and wherein the encoder-decoder network is trained to minimize a loss function that indicates a difference between the generated approximation of the original patch and the original patch. The source models significantly enhance the transfer learning performance for many medical imaging tasks including, but not limited to, disease/organ detection, classification, and segmentation. Other related embodiments are disclosed.
-
公开(公告)号:US20200380695A1
公开(公告)日:2020-12-03
申请号:US16885579
申请日:2020-05-28
Applicant: Zongwei Zhou , Md Mahfuzur Rahman Siddiquee , Nima Tajbakhsh , Jianming Liang
Inventor: Zongwei Zhou , Md Mahfuzur Rahman Siddiquee , Nima Tajbakhsh , Jianming Liang
Abstract: Methods, systems, and media for segmenting images are provided. In some embodiments, the method comprises: generating an aggregate U-Net comprised of a plurality of U-Nets, wherein each U-Net in the plurality of U-Nets has a different depth, wherein each U-Net is comprised of a plurality of nodes Xi,j, wherein i indicates a down-sampling layer the U-Net, and wherein j indicates a convolution layer of the U-Net; training the aggregate U-Net by: for each training sample in a group of training samples, calculating, for each node in the plurality of nodes Xi,j, a feature map xi,j, wherein xi,j is based on a convolution operation performed on a down-sampling of an output from Xi−1,j when j=0, and wherein xi,j is based on a convolution operation performed on an up-sampling operation of an output from Xi+1,j−1 when j>0; and predicting a segmentation of a test image using the trained aggregate U-Net.
-
公开(公告)号:US12216737B2
公开(公告)日:2025-02-04
申请号:US17698805
申请日:2022-03-18
Inventor: Zongwei Zhou , Jae Shin , Jianming Liang
IPC: G06V10/82 , G06F18/21 , G06F18/214 , G06T7/00 , G06V10/764 , G16H30/40
Abstract: Described herein are systems, methods, and apparatuses for actively and continually fine-tuning convolutional neural networks to reduce annotation requirements, in which the trained networks are then utilized in the context of medical imaging. The success of convolutional neural networks (CNNs) in computer vision is largely attributable to the availability of massive annotated datasets, such as ImageNet and Places. However, it is tedious, laborious, and time consuming to create large annotated datasets, and demands costly, specialty-oriented skills. A novel method to naturally integrate active learning and transfer learning (fine-tuning) into a single framework is presented to dramatically reduce annotation cost, starting with a pre-trained CNN to seek “worthy” samples for annotation and gradually enhances the (fine-tuned) CNN via continual fine-tuning. The described method was evaluated using three distinct medical imaging applications, demonstrating that it can reduce annotation efforts by at least half compared with random selection.
-
公开(公告)号:US20250029372A1
公开(公告)日:2025-01-23
申请号:US18776010
申请日:2024-07-17
Inventor: Zuwei Guo , Nahid Ul Islam , Jianming Liang
IPC: G06V10/774 , G06N3/0455 , G06N3/088 , G06N3/094 , G06V10/82 , G16H30/40
Abstract: A stepwise incremental pre-training for integrating discriminative, restorative, and adversarial learning in an AI model. Exemplary systems include means for receiving a training dataset for training a unified AI model. The unified AI model applies each of discriminative, restorative, and adversarial learning operations through three transferable components: a discriminative encoder, a restorative decoder, and an adversarial encoder. Stepwise incremental pre-training operations train the unified AI model, including pre-training the discriminative encoder via discriminative learning and attaching the trained discriminative encoder with the restorative decoder to form a skip-connected encoder-decoder, pre-training the skip-connected encoder-decoder via joint discriminative and restorative learning, and associating the pre-trained skip-connected encoder-decoder with the adversarial encoder. Training of the AI model is finalized by performing full discriminative, restorative, and adversarial learning on the training dataset using the unified AI model.
-
公开(公告)号:US20240339200A1
公开(公告)日:2024-10-10
申请号:US18627831
申请日:2024-04-05
Inventor: DongAo Ma , Jiaxuan Pang , Jianming Liang
IPC: G16H30/40 , G06V10/764 , G16H30/20
CPC classification number: G16H30/40 , G06V10/764 , G16H30/20
Abstract: Exemplary systems include means for receiving medical image data at the system from a plurality of datasets provided via publicly available sources; evaluating the medical image data for the presence of expert notation embedded within the medical image data; determining the expert notations embedded within the medical image data are formatted using inconsistent and heterogeneous labeling across the plurality of datasets; generating an interim AI model by applying a task head classifier to learn the annotations of the expert notations embedded within the medical image data to generate an interim AI model; scaling the interim AI model having the learned annotations of the expert notations embedded therein to additional tasks by applying multi-task heads using cyclical pre-training of the interim AI model trained previously to generate task-specific AI models, with each respective task-specific AI model having differently configured task-specific learning objectives; training a pre-trained AI model specially configured for an application-specific target task by applying task re-visitation training forcing the pre-trained AI model being trained to re-visit all tasks in each round of training and forcing the pre-trained AI model being trained to re-use all accrued knowledge to improve learning by the pre-trained AI model being trained against the current application-specific target task for which the pre-trained AI model is being trained.
-
公开(公告)号:US20240078434A1
公开(公告)日:2024-03-07
申请号:US18241811
申请日:2023-09-01
Inventor: Zuwei Guo , Nahid Ul Islam , Jianming Liang
IPC: G06N3/0895 , G06N3/0455 , G06T7/00 , G06V10/774 , G06V10/82 , G06V20/50 , G06V20/70
CPC classification number: G06N3/0895 , G06N3/0455 , G06T7/0012 , G06V10/774 , G06V10/82 , G06V20/50 , G06V20/70 , G06T2207/20081 , G06T2207/20084 , G06T2207/30004 , G06V2201/03
Abstract: The system receives a plurality of medical images and integrates Self-Supervised machine Learning (SSL) instructions for performing a discriminative learning operation, a restorative learning operation, and an adversarial learning operation into a model for processing the received plurality of medical images. The model is configured with each of a discriminative encoder, a restorative decoder, and an adversarial encoder. Each of the discriminative encoder and the restorative decoder are configured to be skip connected, forming an encoder-decoder. Step-wise incremental training to incrementally train each of the discriminative encoder, the restorative decoder, and the adversarial encoder is performed, in particular: pre-training the discriminative encoder via discriminative learning; attaching the pre-trained discriminative encoder to the restorative decoder to configure the encoder-decoder as a pre-trained encoder-decoder; and training the pre-trained encoder-decoder of the model using joint discriminative and restorative learning. The pre-trained encoder-decoder is associated with the adversarial encoder. The pre-trained encoder-decoder associated with the adversarial encoder is trained through discriminative, restorative, and adversarial learning to render a trained model for the processing of the received plurality of medical images. The plurality of medical images are processed through the model using the trained model.
-
公开(公告)号:US11763952B2
公开(公告)日:2023-09-19
申请号:US17180575
申请日:2021-02-19
Inventor: Fatemeh Haghighi , Mohammad Reza Hosseinzadeh Taher , Zongwei Zhou , Jianming Liang
IPC: G06K9/00 , G16H50/70 , G16H30/40 , G16H30/20 , G06F16/55 , G06N3/08 , G06F16/583 , G06F18/28 , G06F18/214 , G06V10/772 , G06V10/82
CPC classification number: G16H50/70 , G06F16/55 , G06F16/583 , G06F18/214 , G06F18/28 , G06N3/08 , G06V10/772 , G06V10/82 , G16H30/20 , G16H30/40
Abstract: Described herein are means for learning semantics-enriched representations via self-discovery, self-classification, and self-restoration in the context of medical imaging. Embodiments include the training of deep models to learn semantically enriched visual representation by self-discovery, self-classification, and self-restoration of the anatomy underneath medical images, resulting in a collection of semantics-enriched pre-trained models, called Semantic Genesis. Other related embodiments are disclosed.
-
公开(公告)号:US20230081305A1
公开(公告)日:2023-03-16
申请号:US17944881
申请日:2022-09-14
Inventor: Nahid Ul Islam , Shiv Gehlot , Zongwei Zhou , Jianming Liang
Abstract: Described herein are means for systematically determining an optimal approach for the computer-aided diagnosis of a pulmonary embolism, in the context of processing medical imaging. According to a particular embodiment, there is a system specially configured for diagnosing a Pulmonary Embolism (PE) within new medical images which form no part of the dataset upon which the AI model was trained. Such a system executes operations for receiving a plurality of medical images and processing the plurality of medical images by executing an image-level classification algorithm to determine the presence or absence of a Pulmonary Embolism (PE) within each image via operations including: pre-training an AI model through supervised learning to identify ground truth; fine-tuning the pre-trained AI model specifically for PE diagnosis to generate a pre-trained PE diagnosis and detection AI model; wherein the pre-trained AI model is based on a modified CNN architecture having introduced therein a squeeze and excitation (SE) block enabling the CNN architecture to extract informative features from the plurality of medical images by fusing spatial and channel-wise information; applying the pre-trained PE diagnosis and detection AI model to new medical images to render a prediction as to the presence or absence of the Pulmonary Embolism within the new medical images; and outputting the prediction as a PE diagnosis for a medical patient.
-
公开(公告)号:US11436725B2
公开(公告)日:2022-09-06
申请号:US17098422
申请日:2020-11-15
Inventor: Mohammad Reza Hosseinzadeh Taher , Fatemeh Haghighi , Jianming Liang
Abstract: Not only is annotating medical images tedious and time consuming, but it also demands costly, specialty-oriented expertise, which is not easily accessible. To address this challenge, a new self-supervised framework is introduced: TransVW (transferable visual words), exploiting the prowess of transfer learning with convolutional neural networks and the unsupervised nature of visual word extraction with bags of visual words, resulting in an annotation-efficient solution to medical image analysis. TransVW was evaluated using NIH ChestX-ray14 to demonstrate its annotation efficiency. When compared with training from scratch and ImageNet-based transfer learning, TransVW reduces the annotation efforts by 75% and 12%, respectively, in addition to significantly accelerating the convergence speed. More importantly, TransVW sets new records: achieving the best average AUC on all 14 diseases, the best individual AUC scores on 10 diseases, and the second best individual AUC scores on 3 diseases. This performance is unprecedented, because heretofore no self-supervised learning method has outperformed ImageNet-based transfer learning and no annotation reduction has been reported for self-supervised learning. These achievements are contributable to a simple yet powerful observation: The complex and recurring anatomical structures in medical images are natural visual words, which can be automatically extracted, serving as strong yet free supervision signals for CNNs to learn generalizable and transferable image representation via self-supervision.
-
公开(公告)号:US11164067B2
公开(公告)日:2021-11-02
申请号:US16556130
申请日:2019-08-29
Inventor: Jianming Liang , Zongwei Zhou , Md Mahfuzur Rahman Siddiquee , Nima Tajbakhsh
Abstract: Disclosed are provided systems, methods, and apparatuses for implementing a multi-resolution neural network for use with imaging intensive applications including medical imaging. For example, a system having means to execute a neural network model formed from a plurality of layer blocks including an encoder layer block which precedes a plurality of decoder layer blocks includes: means for associating a resolution value with each of the plurality of layer blocks; means for processing via the encoder layer block a respective layer block input including a down-sampled layer block output processing, via decoder layer blocks, a respective layer block input including an up-sampled layer block output and a layer block output of a previous layer block associated with a prior resolution value of a layer block which precedes the respective decoder layer block; and generating the respective layer block output by summing or concatenating the processed layer block inputs.
-
-
-
-
-
-
-
-
-