This PhD thesis investigates the role of eXplainable Artificial Intelligence (XAI) and Convolutional Neural Networks (CNNs) in classifying and localizing various diseases within biomedical imaging. The primary goal is to address the "black box" nature of deep learning models by employing Class Activation Mapping (CAM) algorithms, such as Grad-CAM and Score-CAM, to provide visual heatmaps that justify AI predictions. This transparency is essential for building clinical trust and transforming AI into a reliable "co-pilot" for medical professionals. The research spans three imaging categories: Skin Surface, Internal Organs (e.g., MRI, RX), and Cellular/Tissue Imaging. Additionally, the thesis examines Adversarial Machine Learning to ensure system resilience against malicious attacks and extends XAI methodologies to audio signal classification, including COVID-19 detection from cough recordings. Ultimately, this work establishes a foundation for trustworthy and secure automated diagnostic systems that prioritize both high accuracy and patient safety.

Investigating the Role of Explainable Artificial Intelligence in Biomedical Image Classification, Localization, and Security

DI GIAMMARCO, MARCELLO
2026

Abstract

This PhD thesis investigates the role of eXplainable Artificial Intelligence (XAI) and Convolutional Neural Networks (CNNs) in classifying and localizing various diseases within biomedical imaging. The primary goal is to address the "black box" nature of deep learning models by employing Class Activation Mapping (CAM) algorithms, such as Grad-CAM and Score-CAM, to provide visual heatmaps that justify AI predictions. This transparency is essential for building clinical trust and transforming AI into a reliable "co-pilot" for medical professionals. The research spans three imaging categories: Skin Surface, Internal Organs (e.g., MRI, RX), and Cellular/Tissue Imaging. Additionally, the thesis examines Adversarial Machine Learning to ensure system resilience against malicious attacks and extends XAI methodologies to audio signal classification, including COVID-19 detection from cough recordings. Ultimately, this work establishes a foundation for trustworthy and secure automated diagnostic systems that prioritize both high accuracy and patient safety.
21-apr-2026
Inglese
Artificial intelligence
Biomedical Imaging
Classification
Explainability
Security
Bruno, Raffaele
Mercaldo, Francesco
Martinelli, Fabio
File in questo prodotto:
File Dimensione Formato  
PhD_Thesis.pdf

accesso aperto

Licenza: Creative Commons
Dimensione 28.55 MB
Formato Adobe PDF
28.55 MB Adobe PDF Visualizza/Apri

I documenti in UNITESI sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.14242/365724
Il codice NBN di questa tesi è URN:NBN:IT:UNIPI-365724