Work place: e-Health Center, Universitat Oberta de Catalunya (UOC), Barcelona, Spain
E-mail: fpradosc@uoc.edu
Website: https://orcid.org/0000-0002-7872-0142
Research Interests:
Biography
Ferran Prados is a neuroscientist and medical imaging specialist with expertise in quantitative MRI analysis and neurodegenerative diseases, particularly multiple sclerosis. He holds a PhD in Medical Imaging from the Universitat de Girona (Spain) and has contributed to the development of advanced image processing pipelines for brain and spinal cord analysis. Dr. Ferran Prados has published extensively and collaborated in large multicenter clinical studies. He has experience in both academic research and translational applications of medical imaging. Currently, he is involved in projects integrating AI with medical imaging biomarkers.
By Maider Abad Eusebio Garcia Ferran Prados Jordi Casas-Roma
DOI: https://doi.org/10.5815/ijigsp.2026.02.01, Pub. Date: 8 Apr. 2026
In the rapidly evolving landscape of medical diagnostics, efficient and accurate tools for disease identification are crucial. This study analyzes three convolutional neural network (CNN) architectures—IRV2, ResNet50, and DenseNet121—pre-trained on ImageNet and RadImageNet datasets for respiratory disease diagnosis using chest radiographs. We used over 10,000 chest X-ray images, including COVID-19, pneumonia, and control cases, to train and evaluate these models. RadImageNet-trained models, particularly ResNet50, achieved superior performance with 94.49% accuracy, 93.92% sensitivity, and 95.59% precision compared to ImageNet-trained counterparts, though the improvement was not statistically significant in most cases. To enhance interpretability, we developed a counterfactual-based method generating visual explanations of critical areas influencing diagnostic outcomes. This approach, not requiring access to training data or model internals, identifies image parts that could change the predicted diagnosis if altered. It aids in understanding model reasoning and can correct misclassifications, successfully reclassifying up to 40.91% of previously misclassified images through our masking method. By providing clear, independent visual explanations, our method aims to foster trust in AI-assisted diagnoses among medical professionals. While preliminary results are promising, further validation with medical experts will help confirm the clinical relevance of the highlighted regions. This will strengthen the transparency and interpretability of AI decision-making in healthcare. The visual nature of these explanations offers a valuable tool for interpreting complex medical image classification models and may enhance the synergy between AI systems and human expertise in diagnostic processes.
[...] Read more.Subscribe to receive issue release notifications and newsletters from MECS Press journals