pp. 683-699
S&M3555 Research Paper of Special Issue https://doi.org/10.18494/SAM4685 Published: February 29, 2024 Evaluating Feature Fusion Techniques with Deep Learning Models for Coronavirus Disease 2019 Chest X-ray Sensor Image Identification [PDF] Chih-Ta Yen, Jia-Xian Liao, and Yi-Kai Huang (Received July 2, 2023; Accepted January 29, 2024) Keywords: COVID-19, convolutional neural network, deep learning, chest X-ray (CXR), contrast-limited adaptive histogram equalization (CLAHE), feature fusion
Current diagnostic methods for coronavirus disease 2019 (COVID-19) mainly rely on reverse transcription polymerase chain reaction (RT-PCR). However, RT-PCR is costly and time-consuming. Therefore, an accurate, rapid, and inexpensive screening method must be developed for the diagnosis of COVID-19. In this study, we combined image processing technologies with deep learning algorithms to enhance the accuracy of COVID-19 identification from chest X-ray (CXR) sensor images. Contrast-limited adaptive histogram equalization (CLAHE) was used to improve the visibility level of unclear images. In addition, we examined whether our image fusion technique can effectively improve the performance of seven deep learning models (MobileNetV2, ResNet50, ResNet152V2, Inception-ResNet-v2, DenseNet121, DenseNet201, and Xception). The proposed feature fusion technique involves merging the features of an original image with those of an image subjected to CLAHE and then using the merged features to retrain, test, and validate deep learning models for identifying COVID-19 in CXR images. To avoid incidences of images not matching reality and to ensure high model stability, no data enhancement was conducted. The results of this study indicate that the proposed image fusion technique can improve the classification evaluation indicators, especially the sensitivity of deep learning models in two-class and three-class sortings. Sensitivity refers to a model’s ability to detect an infection correctly. The highest accuracy in this study was achieved when combining Xception with the proposed feature fusion technique. In three-class sorting, the accuracy of this method was 99.74%, with the average accuracy of fivefold cross-validation being 99.19%. In two-class sorting, the accuracy of the aforementioned method was 99.74%, with the average accuracy of fivefold cross-validation being 99.50%. The results showed that the proposed image processing technologies with deep learning algorithms have exceptional generalization.
Corresponding author: Chih-Ta YenThis work is licensed under a Creative Commons Attribution 4.0 International License. Cite this article Chih-Ta Yen, Jia-Xian Liao, and Yi-Kai Huang, Evaluating Feature Fusion Techniques with Deep Learning Models for Coronavirus Disease 2019 Chest X-ray Sensor Image Identification, Sens. Mater., Vol. 36, No. 2, 2024, p. 683-699. |