Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds
Article
Sobahi, Nebras, Atila, Orhan, Deniz, Erkan, Sengur, Abdulkadir and Acharya, U. Rajendra. 2022. "Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds." Biocybernetics and Biomedical Engineering. 42 (3), pp. 1066-1080. https://doi.org/10.1016/j.bbe.2022.08.005
Article Title | Explainable COVID-19 detection using fractal dimension and vision transformer with Grad-CAM on cough sounds |
---|---|
ERA Journal ID | 211872 |
Article Category | Article |
Authors | Sobahi, Nebras, Atila, Orhan, Deniz, Erkan, Sengur, Abdulkadir and Acharya, U. Rajendra |
Journal Title | Biocybernetics and Biomedical Engineering |
Journal Citation | 42 (3), pp. 1066-1080 |
Number of Pages | 15 |
Year | 2022 |
Publisher | Elsevier BV |
Place of Publication | Netherlands |
ISSN | 0208-5216 |
2391-467X | |
Digital Object Identifier (DOI) | https://doi.org/10.1016/j.bbe.2022.08.005 |
Web Address (URL) | https://www.sciencedirect.com/science/article/pii/S020852162200078X |
Abstract | The polymerase chain reaction (PCR) test is not only time-intensive but also a contact method that puts healthcare personnel at risk. Thus, contactless and fast detection tests are more valuable. Cough sound is an important indicator of COVID-19, and in this paper, a novel explainable scheme is developed for cough sound-based COVID-19 detection. In the presented work, the cough sound is initially segmented into overlapping parts, and each segment is labeled as the input audio, which may contain other sounds. The deep Yet Another Mobile Network (YAMNet) model is considered in this work. After labeling, the segments labeled as cough are cropped and concatenated to reconstruct the pure cough sounds. Then, four fractal dimensions (FD) calculation methods are employed to acquire the FD coefficients on the cough sound with an overlapped sliding window that forms a matrix. The constructed matrixes are then used to form the fractal dimension images. Finally, a pretrained vision transformer (ViT) model is used to classify the constructed images into COVID-19, healthy and symptomatic classes. In this work, we demonstrate the performance of the ViT on cough sound-based COVID-19, and a visual explainability of the inner workings of the ViT model is shown. Three publically available cough sound datasets, namely COUGHVID, VIRUFY, and COSWARA, are used in this study. We have obtained 98.45%, 98.15%, and 97.59% accuracy for COUGHVID, VIRUFY, and COSWARA datasets, respectively. Our developed model obtained the highest performance compared to the state-of-the-art methods and is ready to be tested in real-world applications. |
Keywords | Cough sound; Fractal dimension; YAMNet; COVID-19 detection; Vision Transformer |
ANZSRC Field of Research 2020 | 400306. Computational physiology |
Public Notes | Files associated with this item cannot be displayed due to copyright restrictions. |
Byline Affiliations | King Abdulaziz University, Saudi Arabia |
Firat University, Turkey | |
Ngee Ann Polytechnic, Singapore | |
Asia University, Taiwan | |
Singapore University of Social Sciences (SUSS), Singapore |
Permalink -
https://research.usq.edu.au/item/z1w28/explainable-covid-19-detection-using-fractal-dimension-and-vision-transformer-with-grad-cam-on-cough-sounds
49
total views0
total downloads3
views this month0
downloads this month