This short paper reports about a line of research exploiting a conditional logic of commonsense reasoning to provide a semantic interpretation to neural network models. A “concept-wise" multi-preferential semantics for conditionals is exploited to build a preferential interpretation of a trained neural network starting from its input-output behavior. The approach is general (model agnostic): it is based on a notion of metric distance to define preferences and has been first proposed for Self-Organising Maps (SOMs). For MultiLayer Perceptrons (MLPs), a deep network can as well be regarded as a (fuzzy) conditional knowledge base (KB), in which the synaptic connections correspond to weighted conditionals. This opens to the possibility of adopting conditional description logics as a basis for neuro-symbolic integration. Proof methods for many-valued weighted conditional KBs have been developed, based on Answer Set Programming and Datalog encodings to deal with the entailment and model-checking problems.

From Common Sense Reasonig to Neural Network Models: a Conditional and Multi-preferential Approach for Explainability and Neuro-Symbolic Integration

Marco Botta;Roberto Esposito;Valentina Gliozzi;
2022-01-01

Abstract

This short paper reports about a line of research exploiting a conditional logic of commonsense reasoning to provide a semantic interpretation to neural network models. A “concept-wise" multi-preferential semantics for conditionals is exploited to build a preferential interpretation of a trained neural network starting from its input-output behavior. The approach is general (model agnostic): it is based on a notion of metric distance to define preferences and has been first proposed for Self-Organising Maps (SOMs). For MultiLayer Perceptrons (MLPs), a deep network can as well be regarded as a (fuzzy) conditional knowledge base (KB), in which the synaptic connections correspond to weighted conditionals. This opens to the possibility of adopting conditional description logics as a basis for neuro-symbolic integration. Proof methods for many-valued weighted conditional KBs have been developed, based on Answer Set Programming and Datalog encodings to deal with the entailment and model-checking problems.
2022
8th Workshop on Formal and Cognitive Reasoning co-located with the 45th German Conference on Artificial Intelligence (KI 2022)
Trier
19/9/2022
Proceedings of the 8th Workshop on Formal and Cognitive Reasoning co-located with the 45th German Conference on Artificial Intelligence (KI 2022)
CEUR
3242
66
78
https://ceur-ws.org/Vol-3242/paper5.pdf
Preferential Description Logics, Typicality, Neural Networks, Explainability
Francesco Bartoli, Marco Botta, Roberto Esposito, Laura Giordano, Valentina Gliozzi ,Daniele Theseider Dupré
File in questo prodotto:
File Dimensione Formato  
paper5.pdf

Accesso aperto

Dimensione 280.91 kB
Formato Adobe PDF
280.91 kB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2318/1892058
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? ND
social impact