Modern Machine Learning (ML) has significantly advanced various research fields, but the opaque nature of ML models hinders their adoption in several domains. Explainable AI (XAI) addresses this challenge by providing additional information to help users understand the internal decision-making process of ML models. In the field of neuroscience, enriching a ML model for brain decoding with attribution-based XAI techniques means being able to highlight which brain areas correlate with the task at hand, thus offering valuable insights to domain experts. In this paper, we analyze human and Computer Vision (CV) systems in parallel, training and explaining two ML models based respectively on functional Magnetic Resonance Imaging (fMRI) and movie frames. We do so by leveraging the “StudyForrest” dataset, which includes functional Magnetic Resonance Imaging (fMRI) scans of subjects watching the “Forrest Gump” movie, emotion annotations, and eye-tracking data. For human vision the ML task is to link fMRI data with emotional annotations, and the explanations highlight the brain regions strongly correlated with the label. On the other hand, for computer vision, the input data is movie frames, and the explanations are pixel-level heatmaps. We cross-analyzed our results, linking human attention (obtained through eye-tracking) with XAI saliency on CV models and brain region activations. We show how a parallel analysis of human and computer vision can provide useful information for both the neuroscience community (allocation theory) and the ML community (biological plausibility of convolutional models)

Explainable Emotion Decoding for Human and Computer Vision

Borriero A.
First
Membro del Collaboration Group
;
Diano M.
Membro del Collaboration Group
;
Orsenigo D.
Membro del Collaboration Group
;
Tamietto M.
Membro del Collaboration Group
;
2024-01-01

Abstract

Modern Machine Learning (ML) has significantly advanced various research fields, but the opaque nature of ML models hinders their adoption in several domains. Explainable AI (XAI) addresses this challenge by providing additional information to help users understand the internal decision-making process of ML models. In the field of neuroscience, enriching a ML model for brain decoding with attribution-based XAI techniques means being able to highlight which brain areas correlate with the task at hand, thus offering valuable insights to domain experts. In this paper, we analyze human and Computer Vision (CV) systems in parallel, training and explaining two ML models based respectively on functional Magnetic Resonance Imaging (fMRI) and movie frames. We do so by leveraging the “StudyForrest” dataset, which includes functional Magnetic Resonance Imaging (fMRI) scans of subjects watching the “Forrest Gump” movie, emotion annotations, and eye-tracking data. For human vision the ML task is to link fMRI data with emotional annotations, and the explanations highlight the brain regions strongly correlated with the label. On the other hand, for computer vision, the input data is movie frames, and the explanations are pixel-level heatmaps. We cross-analyzed our results, linking human attention (obtained through eye-tracking) with XAI saliency on CV models and brain region activations. We show how a parallel analysis of human and computer vision can provide useful information for both the neuroscience community (allocation theory) and the ML community (biological plausibility of convolutional models)
2024
2nd World Conference on Explainable Artificial Intelligence, xAI 2024
Valletta
17-19 Luglio 2024
Communications in Computer and Information Science
Springer Science and Business Media Deutschland GmbH
2154
178
201
https://www.springerprofessional.de/en/explainable-emotion-decoding-for-human-and-computer-vision/27331926
Computer Vision; Emotion Recognition; eXplainable Artificial Intelligence; NeuroImaging; Neuroscience
Borriero, A., Milzzo, M., Diano, M., Orsenigo, D., Villa, M.C., DiFazio, C., Tamietto, M., Perotti, A.
File in questo prodotto:
File Dimensione Formato  
final_8274.pdf

Accesso aperto

Descrizione: Paper
Tipo di file: PDF EDITORIALE
Dimensione 5.55 MB
Formato Adobe PDF
5.55 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2318/2018250
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact