Research Grant | May 2023
Project Title: Evaluation of New AI Methods for Ultrasound Imaging Based Subclinical Atherosclerosis Risk Assessment
Funding Agency: The Heart Foundation of Northern Sweden
Funding Amount: 200000 SEK
Principal Investigator: Mohd Usama,
Co- Investigator: Assoc Prof. Christer Gr ¨onlund
Description: Ultrasound imaging of the carotid arteries is an important method for the assessment of subclinical atherosclerosis. In atherosclerotic disease, fatty streaks start infiltrating the arterial walls and may eventually build up to localized plaques, which may rupture and cause stroke. However, ultrasound images are difficult to read because of operator dependence and noise contamination. In this study, artificial intelligence will be used to develop new decision support systems aiming to overcome some of these difficulties.
Travel Grant | March 2023
Purpose: To attend conferences and invited talk.
Funding Agency: JC Kempes Memorial Academic Research Fund, Sweden
Funding Amount: 11205 SEK
Domain Adaptation of Carotid Ultrasound
Deep learning has been used extensively for medical image analysis applications, assuming the training and test data adhere to the same probability distributions. However, a common challenge arises when dealing with medical images generated by different systems or even the same system with varying parameter settings. Such images often contain diverse textures and noise patterns, violating the assumption. Consequently, models trained on data from one machine or setting usually struggle to perform effectively on data from another. To address this issue in ultrasound images, we proposed a Generative Adversarial Network (GAN) based model in this paper. We formulated image harmonization and denoising tasks as an image-to-image translation task, wherein we modified the texture pattern and reduced noise in Carotid ultrasound images while keeping the image content (the anatomy) unchanged. The performance was evaluated using feature distribution and pixel-space similarity metrics. In addition, blood-to-tissue contrast and influence on computed risk markers (Gray scale median, GSM) were evaluated. The results showed that domain adaptation was achieved in both tasks (histogram correlation 0.920 and 0.844), as compared to no adaptation (0.890 and 0.707), and that the anatomy of the images was retained (structure similarity index measure of the arterial wall 0.71 and 0.80). In addition, the image noise level (contrast) did not change in the image harmonization task (-34.1 vs 35.2 dB) but was improved in the noise reduction task (-23.5 vs -46.7 dB). The model outperformed the CycleGAN in both tasks. Finally, the risk marker GSM increased by 7.6 (p<0.001) in task 1 but not in task 2. We conclude that domain translation models are powerful tools for ultrasound image improvement while retaining the underlying anatomy but that downstream calculations of risk markers may be affected.
ArXiv, 2024
https://arxiv.org/abs/2407.05163
Carotid Ultrasound Image Denoising
The Medical Technology Days (MTD) 2023, Stockholm
https://www.medicinteknikdagarna.se/wp-content/uploads/sites/17/2023/09/MTD-2023-abstractbok-1.pdf#page=100
Ultrasound images play a crucial role in the diagnosis and treatment of various diseases, providing clinicians with valuable insights into a patient's condition. However, ultrasound images are often subjected to various challenges, such as noise, artifacts, and limited resolution, which can hinder accurate analysis and interpretation. In recent years, generative adversarial networks (GANs) have emerged as a promising approach for enhancing the quality of medical images. Despite the advancements made, challenges remain in the application of GANs for ultrasound image quality enhancement. In this paper, we formulate ultrasound image denoising task as a domain-to-domain translation task between images of low and high quality. In this way we can translate a noisy image into high quality image while the anatomical content of the image is unchanged. We propose a finetuned Cycle-GAN objective function by adding a noise loss, which reduces the noise from images by minimizing the difference among features from early three layers of generator. The generator aims to generate high quality images from noisy images by transferring the noise to another domain, and the discriminator enforces the generated images to have the same content as original images. Several models (CycleGAN, BiGAN, DualGAN and our proposed cycleGAN) were trained on 500 carotid images of each domain, and translation performance was quantified using histogram correlation distances. Results show that the method is able to translate the noisy low-quality images into a high quality images. The histogram correlation distances between the low-quality test images and the corresponding translated images were 0.96, 0.71, 0.60 and 0.85 for the proposedGAN, biGAN, cycleGAN and dualGAN, respectively. The experimental results show that the proposed method creates robust transferable features between two domains, and improves the denoising performance compared to the state-of-art methods. The method may be useful for restoration/standardization of retrospective low-quality ultrasound image sequences.
Multimodal Disease Risk Assessment
IEEE Access
https://ieeexplore.ieee.org/abstract/document/8519726/
This paper presents the analysis of real-life medical big data obtained from a hospital in central China from 2013 to 2015 for risk assessment of cerebral infarction disease. We propose a new recurrent convolutional neural network (RCNN)-based disease risk assessment multimodel by utilizing structured and unstructured text data from the hospital. In the proposed model, the convolutional layer becomes a bidirectional recurrent neural network by utilizing the intra-layer recurrent connection within the convolutional layer. Each neuron within convolutional layer receives feedforward and recurrent inputs from the previous unit and neighborhood, respectively. In addition to step-by-step recurrent operation, the region of context capture increases, thereby facilitating fine-grain feature extraction. Furthermore, we use a data parallelism approach over multimodel data during training and testing of the proposed model. Results show that the data parallelism approach leads to fast conversion speed. The RCNN-based model works differently from the traditional convolutional neural network and other typical methods. The proposed model exhibits a prediction accuracy of 96.02%, which is higher than those of typical existing methods.
Disease Prediction Using Clinical Notes/Text
https://www.sciencedirect.com/science/article/abs/pii/S0169260719311708
Nowadays computer-aided disease diagnosis from medical data through deep learning methods has become a wide area of research. Existing works of analyzing clinical text data in the medical domain, which substantiate useful information related to patients with disease in large quantity, benefits early-stage disease diagnosis. However, benefits of analysis not achieved well when the traditional rule-based and classical machine learning methods used; which are unable to handle the unstructured clinical text and only a single method is not able to handle all challenges related to the analysis of the unstructured text, Moreover, the contribution of all words in clinical text is not the same in the prediction of disease. Therefore, there is a need to develop a neural model which solve the above clinical application problems, is an interesting topic which needs to be explored. Thus considering the above problems, first, this paper present self-attention based recurrent convolutional neural network (RCNN) model using real-life clinical text data collected from a hospital in Wuhan, China. This model automatically learns high-level semantic features from clinical text by using bi-direction recurrent connection within convolution. Second, to deal with other clinical text challenges, we combine the ability of RCNN with the self-attention mechanism. Thus, self-attention gets the focus of the model on essential convolve features which have effective meaning in the clinical text by calculating the probability of each convolve feature through softmax. The proposed model is evaluated on real-life hospital dataset and used measurement metrics as Accuracy and recall. Experiment results exhibit that the proposed model reaches up to accuracy 95.71%, which is better than many existing methods for cerebral infarction disease. This article presented the self-attention based RCNN model by combining the RCNN with self-attention mechanism for prediction of cerebral infarction disease. The obtained results show that the presented model better predict the cerebral infarction disease risk compared to many existing methods. The same model can also be used for the prediction of other disease risks.
Infant Brain Segmentation
https://www.sciencedirect.com/science/article/abs/pii/S0167739X18332291
Magnetic Resonance Imaging (MRI) is dominant modality for infant brain analysis. Segmenting the whole infant MRI brain into number of tissues such as Cerebrospinal fluid (CSF), White matter (WM), and Gray Matter (GM) are highly desirable in the clinical environment. However, traditional methods tend to be degrading due to low contrast between GM and WM in isointense phase (about 6–8 months of early life). Recently, Convolutional Neural Network (CNN) emerged as a robust intelligent approach to examine medical image. The UNet model is among the preferred CNN models that have been widely used for medical imaging applications and achieved excellent results. The UNet architecture is a combination of convolutional, pooling, and up-sampling layers. Recently, 3D-UNet architecture used to exploit 3D-contextual information of volumetric data in many applications. However, CNN faces challenge to distinguish the similar brain tissues. In this paper, we present a variant of 3D-UNet to extract the volumetric contextual information of medical data. We propose a novel combined architecture of dense connection, residual connection, and inception module. The proposed architecture contains three stages, namely the densely connected stage, a residual inception stage, and an up-sampling stage. Our proposed approach provides state-of-art results in comparison to other existing approaches. This suggested approach achieves dice scores of 0.95, 0.905, and 0.92 in CSF, WM, and GM tissues respectively.
Skin Disease Classification
ArXiv
https://arxiv.org/abs/2406.00696
In this study, we proposed a model for skin disease classification using a Bilinear Convolutional Neural Network (BCNN) with a Constrained Triplet Network (CTN). BCNN can capture rich spatial interactions between features in image data. This computes the outer product of feature vectors from two different CNNs by a bilinear pooling. The resulting features encode second-order statistics, enabling the network to capture more complex relationships between different channels and spatial locations. The CTN employs the Triplet Loss Function (TLF) by using a new loss layer that is added at the end of the architecture called the Constrained Triplet Loss (CTL) layer. This is done to obtain two significant learning objectives: inter-class categorization and intra-class concentration with their deep features as often as possible, which can be effective for skin disease classification. The proposed model is trained to extract the intra-class features from a deep network and accordingly increases the distance between these features, improving the model’s performance. The model achieved a mean accuracy of 93.72%.
Sentiment Analysis Using Text Data
Future Generation Computer Systems
https://www.sciencedirect.com/science/article/abs/pii/S0167739X19334600
Convolution and recurrent neural network have obtained remarkable performance in natural language processing(NLP). Moreover, from the attention mechanism perspective convolution neural network(CNN) is applied less than recurrent neural network(RNN). Because RNN can learn long-term dependencies and gives better results than CNN. But CNN has its own advantage, can extract high-level features by using its local fix size context at the input level. Thus, this paper proposed a new model based on RNN with CNN-based attention mechanism by using the merits of both architectures together in one model. In the proposed model, first, CNN learns the high-level features of sentence from input representation. Second, we used attention mechanism to get the attention of the model on the features which contribute much in the prediction task by calculating the attention score from features context generated from CNN filters. Finally, these features context from CNN with attention score are commonly used at the RNN to process them sequentially. To validate the model we experiment on three benchmark datasets. Experiment results and their analysis demonstrate the effectiveness of the model.
1) Disease Prediction Using Multimodal Healthcare Data (Sept 2019-Aug 2021)
Deanship of Scientific Research funded this project at King Saud University, Riyadh, Saudi Arabia, through the Research Group Project under grant number RGP-049.
Description: In this project, we worked on multimodal medical data from 2013 to 2015, consisting of structured and unstructured data, to address the following problem in disease prediction applications. 1) How to extract main characteristics usually not mentioned in patient medical records that affect chronic diseases for risk assessment. 2) Difficulty extracting fine-grain disease-related features from unstructured text data using traditional CNN? These phenomena limit the accurate prediction of disease risk assessment.
2) Skin Cancer Detection Using Deep Learning Over Dermatology Images (Jan 2019-Feb 2020)
This work was funded at the AIRS Institute (Artificial Intelligence and Robotics for Society), Chinese University of Hong Kong, Shenzhen, China under the grant number 501100004853.
Description: Since hand-crafted feature learning algorithms are usually committed to a single or a minimal number of subcategories, they are ineffective in classifying a higher number of skin disease categories and impractical due to variations like skin diseases. To address this issue, more focus is required on feature learning rather than feature engineering to select essential features by machine. To deal with the problem, this project established a discriminative feature learning approach based on transfer-learning for skin disease diagnosis using a new deep CNN with triplet loss function.
M. Usama, E. Nyman, U. Naslund, et al. ”A Domain Adaptation Model for Carotid Ultrasound: Image Harmonization, Noise Reduction, and Impact on Cardiovascular Risk Markers.” Computers in Biology and Medicine, Volume 190, 2025, 110030, ISSN 0010-4825, 2024. (IF 7.0)
M. Usama, B. Ahmad, M. Chen, et al. “Self-attention based recurrent convolutional neural network for disease prediction using healthcare data,” Computer Methods and Programs in Biomedicine, Vol. 190, pp. 0169-2607, 2019. (IF 6.1)
M. Usama, B. Ahmad, J. Wan, et al. “Deep feature learning for disease risk assessment based on convolutional neural network with intra-layer recurrent connection by using hospital big data,” in IEEE Access, vol. 6, pp. 67927-67939, 2018. (IF 3.9)
M. Usama, B. Ahmad, E. Song, et al. “Attention-based sentiment analysis using convolutional and recurrent neural network,” Future Generation Computer Systems, vol. 113, pp. 571-578, 2020. (IF 7.5)
M. Usama, W. Xiao, B. Ahmad, et al. ”Deep learning based weighted feature fusion approach for sentiment analysis,” in IEEE Access, vol. 7, pp. 140252-140260, 2019. (IF 3.9)
M. Usama, M. Liu, and M. Chen, ”Job schedulers for Big data processing in Hadoop environment: testing real-life schedulers using benchmark programs,” Digital Communications and Networks, 3(4), 260273. (IF 7.9)
B. Ahmad, M. Usama, C. Huang, et al. “Discriminative Feature Learning for Skin Disease Classification Using Deep Convolutional Neural Network,” in IEEE Access, vol. 8, pp. 39025-39033, 2020. (IF 3.9)
B. Ahmad, M. Usama, T. Ahmad, et al. ”Bilinear-Convolutional Neural Network Using a Matrix Similarity-based Joint Loss Function for Skin Disease Classification.” arXiv preprint arXiv:2406.00696, 2024. (Preprint)
B. Ahmad, M. Usama, T. Ahmad, et al. ”DACB-Net: Dual Attention Guided Compact Bilinear Convolution Neural Network for Skin Disease Classification.” arXiv preprint arXiv:2407.03439, 2024. (Preprint)
Y. Hao, M. Usama, J. Yang, M. S. Hossain, and A. Ghoneim, ”Recurrent convolutional neural network based multimodal disease risk prediction,” Future Generation Computer Systems, vol. 92, pp. 76-83, 2019. (IF 7.5)
B. Ahmad, M. Usama, T. Ahmad, et al. “An ensemble model of convolution and recurrent neural network for skin disease classification,” International Journal of Imaging Systems and Technology. 2022; 1- 12. (IF 3.3)
S. Qamar, H. Jin, M. Usama, et al. “A variant form of 3D-UNet for infant brain segmentation.” Future Generation Computer Systems, Volume 108, 2020, Pages 613-623, ISSN 0167-739X. (IF 7.5)
M. Usama, B. Ahmad, A. P. Singh, P. Ahmad. “Recurrent convolutional attention neural model for sentiment classification of short text.”IEEE International Conference on Cutting-edge Technologies in Engineering, India 2019. (Scopus)
B. Ahmad, M. Usama, J. Lu et al. ”Deep Convolutional Neural Network Using Triplet Loss to Distinguish the Identical Twins,” 2019 IEEE Globecom Workshops, Waikoloa, HI, USA, 2019. (Scopus)
P. Ahmad, H. Jin, M. Usama et al. ”3D dense dilated hierarchical architecture for brain tumor segmentation.” Proceedings of the 4th International Conference on Big Data and Computing. 2019.(Scopus)
M. Usama, A. Saboori C. Gr ¨onlund, “Carotid ultrasound image denoising using low-to-high image quality domain adaptation,” Medical Technology Days 2023, Stockholm, Sweden. (Poster Presentation)