Volume 17 , Issue 2 , PP: 186-196, 2025 | Cite this article as | XML | Html | PDF | Full Length Article
Vathana D. 1 , Babu S. 2 *
Doi: https://doi.org/10.54216/FPA.170214
Advanced imaging in medical has become crucial in the early identify diseases because they reveal the important structural features of the human body. But it is almost impossible to get such high resolution images in real life situation due to the factors such as image capture and processing equipment, and environmental factors that affect the outcome of the image. This work proposes a sub-type of GAN that is used in enhancement of images particularly in medical fields. The generator of the Med-GAN extracts a high-resolution image from a low-resolution one with the help of novel features learned by the model. The approach of reconstructing high resolution from multiple parallel streams of lower resolution employs deconvolution algorithms with multiple scale fusions that produce better high resolution representations as compared to the technique of bilinear interpolation. The performances of the proposed Med-GAN are tested on two publicly available COVID-19 CT datasets and one private medical image dataset which shows that the proposed method outperforms the existing methods in performance comparisons. Consequently, for PSNR, the score improves from 24.103 dB corresponding to the Initial Approach of the “BRaTS (FLAIR)” dataset to 25.496 dB for the Proposed Method; whereas for SSIM the score increases from 0.782 to 0.812.se types of high-resolution images are usually impossible to get due to limits in imaging devices, environmental conditions, and human factors. This work proposes the Med-GAN: an Enhanced Super-Resolution Generative Adversarial Network tuned for medical image enhancement. The Med-GAN generator learns high-resolution representations from low-resolution images via advanced feature extraction methods. Deconvolution algorithms with multi-scale fusions recover better high-resolution representations from multiple parallel streams of lower resolutions in this approach compared to traditional bilinear interpolation methods. Evaluated on two publicly available COVID-19 CT datasets and one custom medical image dataset, the proposed Med-GAN significantly outperforms existing techniques in performance comparisons. In particular, PSNR rises from 24.103 dB for the "BRaTS (FLAIR)" dataset in the Initial Approach to 25.496 dB in the Proposed Method, while SSIM increases from 0.782 to 0.812. If that is the case then it could be said that the solution of the proposed Med-GAN is one of the most realistic means for improving the quality of medical images and therefore contributes to better diagnostics of diseases
High-resolution medical imaging , Image augmentation , Enhanced Super-Resolution Generative Adversarial Networks , Med-GAN , Super-resolution
[1] Tsuneki M (2022) Deep learning models in medical image analysis. J Oral Biosci. 2022 Sep;64(3):312-320.
[2] Van der Velden BH, Kuijf HJ, Gilhuijs KG, Viergever MA (2022) Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Med Image Anal.
[3] Chen X, Wang X, Zhang K, Fung KM, Thai TC, Moore K, Mannel RS, Liu H, Zheng B, Qiu Y (2022a) Recent advances and clinical applications of deep learning in medical image analysis. Med Image Anal.
[4] Yu H, Yang LT, Zhang Q, Armstrong D, Deen MJ (2021) Convolutional neural networks for medical image analysis: state-of-the-art, comparisons, improvement and perspectives. Neurocomputing 444:92–110.
[5] C. Dong, C. C. Loy, K. He, and X. Tang, “Learning a deep convolutional network for image super-resolution,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 184–199, Springer, Zurich, Switzerland, September 2014.
[6] J. Kim, J. K. Lee, and K. M. Lee, "Accurate image super-resolution using intense convolutional networks," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1646–1654, IEEE, San Juan, PR, USA, June 2016.
[7] Y. Tai, J. Yang, and X. M. Liu, "Image super-resolution via a deep recursive residual network," in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3147–3155, IEEE, Honolulu, HI, USA, July 2017.
[8] Y. L. Zhang, Y. P. Tian, Y. Kong, and B. N. Zhong, “Residual dense network for image super-resolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018.
[9] Z. Hui, X. M. Wang, and X. B. Gao, “Fast and accurate single image super-resolution via information distillation network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 723–731, IEEE, Salt Lake City, UT, USA, June 2018.
[10] J. Li, F. Fang, K. Mei, and G. Zhang, “Multi-scale residual network for image super-resolution,” in Proceedings of the European Conference on Computer Vision (ECCV), pp. 527–542, Springer, Munich, Germany, September 2018.
[11] Z. Li, J. L. Yang, Z. Liu, X. M. Yang, G. Jeon, and W. Wu, “Feedback network for image super-resolution,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3867–3876, IEEE, Long Beach, CA,USA, June 2019.
[12] I. J. Goodfellow, J. P. Abadie, and M. Mirza, “Generative adversarial nets,” in Proceedings of the Annual Conference on Neural Information Processing Systems (ICONIP), pp. 2672–2680, Curran Associates, Montreal, Canada, December 2014.
[13] C. Ledig, L. Theis, F. Husz'ar, A. Cunningham, and A. Acosta, “Photo-realistic single image super-resolution using a generative adversarial network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 105–114, IEEE, Honolulu, HI, USA, July 2017
[14] M. Arjovsky, S. Chintala, and L. Bottou, “Wasserstein GAN,” 2017, https://arxiv.org/abs/1701.07875.
[15] P. Isola, J. Y. Zhu, and T. Zhou, “Image-to-image translation with conditional adversarial networks,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125–1134, IEEE, Honolulu, HI, USA, July 2017.
[16] H. Gao, Z. Chen, B. Huang, J. Chen, and Z. Li, "Image super-resolution based on conditional generative adversarial network," IET Image Processing, vol. 14, no. 13, pp. 3006–3013,2020
[17] X. L. Zun, H. J. Zhong, and L. R. Xing, “Multi-scale generative adversarial networks for image super-resolution algorithms,” Science Technology and Engineering, vol. 20, no. 13, pp. 5217–5223, 2020.
[18] Khan AR, Khan S, Harouni M, Abbasi R, Iqbal S, Mehmood Z (2021) Brain tumour segmentation using K-means clustering and deep learning with synthetic data augmentation for classification. Microsc Res Tech 84:1389–1399
[19] Dufumier B, Gori P, Battaglia I, Victor J, Grigis A, Duchesnay E (2021) Benchmarking CNN on 3d anatomical brain MRI: architectures, data augmentation and deep ensemble learning. arXiv preprint, pp 1–25.
[20] Isensee F, J¨ager PF, Full PM et al (2020) nnu-Net for brain tumor segmentation. In: International MICCAI brainlesion workshop. Springer, Cham, pp 118–132
[21] Fidon L, Ourselin S, Vercauteren T (2020) Generalized wasserstein dice score, distributionally robust deep learning, and ranger for brain tumor segmentation: brats 2020 challenge. in: International MICCAI brain lesion workshop, Lima, Peru, pp 200–214
[22] Wang Y, Ji Y, Xiao H (2022) A Data Augmentation Method for Fully Automatic Brain Tumor Segmentation. arXiv preprint, pp 1–15
[23] Kossen T, Subramaniam P, Madai VI, Hennemuth A, Hildebrand K, Hilbert A, Sobesky J, Livne M, Galinovic I, Khalil AA, Fiebach JB (2021) Synthesizing anonymized and labelled TOF-MRA patches for brain vessel segmentation using generative adversarial networks. Comput Biol Med 131:1–9
[24] Hu R, Ruan G, Xiang S, Huang M, Liang Q, Li J (2020) Automated diagnosis of COVID-19 using deep learning and data augmentation on chest CT. medRxiv, pp 1–11
[25] Alshazly H, Linse C, Barth E et al (2021) Explainable covid-19 detection using chest ct scans and deep learning. Sensors 21:1–22
[26] Wang Q, Zhang X, Zhang W, Gao M, Huang S, Wang J, Zhang J, Yang D, Liu C (2021) Realistic lung nodule synthesis with multi-target co-guided adversarial mechanism. IEEE Trans Med Imaging 40:2343–2353
[27] Nishio M, Muramatsu C, Noguchi S, Nakai H, Fujimoto K, Sakamoto R, Fujita H (2020) Attribute-guided image generation of three-dimensional computed tomography images of lung nodules using a generative adversarial network. Comput Biol Med
[28] Karthiga R, Narasimhan K, Amirtharajan R (2022) Diagnosis of breast cancer for modern mammography using artificial intelligence. Math Comput Simul 202:316–330
[29] Zeiser FA, da Costa CA, Zonta T et al. (2020) Segmentation of masses on mammograms using data augmentation and deep learning. J Digit Imaging 33:858–868
[30] Alyafi B, Diaz O, Marti R (2020) DCGANs for realistic breast mass augmentation in X-ray mammography. IN: Medical imaging 2020: computer-aided diagnosis, International Society for Optics and Photonics, pp 1–4
[31] Shen T, Hao K, Gou C, Wang FY (2021) Mass image synthesis in mammogram with contextual information based on GANS. Comput Methods Programs Biomed
[32] Shyamalee T, Meedeniya (2022) D CNN based fundus images classification for glaucoma identification. In: 2nd International conference on advanced research in computing (ICARC), Belihuloya, Sri Lanka, pp 200–205
[33] Tufail AB, Ullah I, Khan WU, Asif M, Ahmad I, Ma YK, Khan R, Ali M (2021) Diagnosis of diabetic retinopathy through retinal fundus images and 3D convolutional neural networks with a limited number of samples. Wirel Commun Mob Comput 2021:1–15
[34] Kurup A, Soliz P, Nemeth S, Joshi V (2020) Automated detection of malarial retinopathy using transfer learning. In: IEEE Southwest Symposium on Image Analysis and Interpretation (SSIAI), Albuquerque, USA, pp 18–21
[35] Sun X, Fang H, Yang Y et al. (2021) Robust retinal vessel segmentation from a data augmentation perspective. In: International workshop on ophthalmic medical image analysis, pp 189–198
[36] Agustin T, Utami E, Al Fatta H (2020) Implementation of data augmentation to improve performance CNN method for detecting diabetic retinopathy. In: 3rd International Conference on information and Communications Technology (ICOIACT), Indonesia, Yogyakarta, pp 83–88
[37] Zhou Y, Wang B, He X, Cui S, Shao L (2020) DR-GAN: conditional generative adversarial network for fine-grained lesion synthesis on diabetic retinopathy images. IEEE J Biomedical Health Inf 26:56–66
[38] Balasubramanian R, Sowmya V, Gopalakrishnan EA, Menon VK, Variyar VS, Soman KP (2020) Analysis of adversarial based augmentation for diabetic retinopathy disease grading. In: 11th International Conference on Computing, Communication and Networking Technologies (ICT), India, Kharagpur, pp 1–5
[39] Araújo T, Aresta G, Mendonça L et al. (2020) Data augmentation for improving proliferative diabetic retinopathy detection in eye fundus images. IEEE Access 8:462–474
[40] Bakas S, Reyes M, Jakab A, Bauer S, Rempfler M, Crimi A, Shinohara RT, Berger C, Ha SM, Rozycki M, Prastawa M (2018) Identifying the best machine learning algorithms for brain tumour segmentation, progression assessment, and overall survival prediction in the brats challenge. arXiv preprint, pp 1–49
[41] Armato SG, McLennan G, Bidaut L, McNitt-Gray MF, Meyer CR, Reeves AP, Zhao B, Aberle DR, Henschke CI, Hoffman EA, Kazerooni EA (2011) The lung image database consortium (LIDC) and image database resource initiative (IDRI): a completed reference database of lung nodules on CT scans. Med Phys 38:915–931
[42] Wang W, Luo J, Yang X, Lin H (2015) Data analysis of the lung imaging database consortium and image database resource initiative. Acad Radiol 22:488–495
[43] Setio AA, Traverso A, De Bel T, Berens MS, Van Den Bogaard C, Cerello P, Chen H, Dou Q, Fantacci ME, Geurts B, van der Gugten R (2017) Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in computed tomography images: the luna16 challenge. Med Image Anal 42:1–3
[44] Manos D, Seely JM, Taylor J, Borgaonkar J, Roberts HC, Mayo JR (2014) The lung reporting and data system (LU-RADS): a proposal for computed tomography screening. Can Assoc Radiol J 65:121–134
[45] Naidich DP, Bankier AA, MacMahon H, Schaefer-Prokop CM, Pistolesi M, Goo JM, Macchiarini P, Crapo JD, Herold CJ, Austin JH, Travis WD (2013) Recommendations for the management of subsolid pulmonary nodules detected at CT: a statement from the Fleischner Society. Radiology 266:304–317
[46] Moreira IC, Amaral I, Domingues I, Cardoso A, Cardoso MJ, Cardoso JS (2012) Inbreast: toward a full-field digital mammographic database. Acad Radiol 19:236–248
[47] Decencière E, Zhang X, Cazuguel G, Lay B, Cochener B, Trone C, Gain P, Ordonez R, Massin P, Erginay A et al (2014) Feedback on a publicly distributed image database: the messidor database. Image Anal Stereol 33:231–234.
[48] Ledig, C., Theis, L., Husz´ar, F., Caballero, J., Cunningham, A., Acosta, A., Aitken, A., Tejani, A., Totz, J., Wang, Z., et al.: Photo-realistic single image superresolution using a generative adversarial network. In: CVPR (2017)
[49] Zhang, Y., Tian, Y., Kong, Y., Zhong, B., Fu, Y.: Residual dense network for image super-resolution. In: CVPR (2018)
[50] S. H. Gao, M. M. Cheng, K. Zhao, X. Y. Zhang, M. H. Yang, and P. Torr, “Res2Net: a new multi-scale backbone architecture," IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI), IEEE, vol. 43, 2020
[51] K.Sun, B. Xiao, D. Liu, and J. D. Wang, “Deep high-resolution representation learning for human pose estimation,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019