In this paper, we propose a new theoretical approach to Explainable AI. Following the Scientific Method, this approach consists of formulating, on the basis of empirical evidence, a mathematical model to explain and predict the behaviors of Neural Networks. We apply the method to a case study created in a controlled environment, which we call Prime Convolutional Model (p-Conv for short). p-Conv operates on a dataset consisting of the first one million natural numbers and is trained to identify the congruence classes modulo a given integer m. Its architecture uses a convolutional-type neural network that contextually processes a sequence of B consecutive numbers for each input. We take an empirical approach and exploit p-Conv to identify the congruence classes of numbers in a validation set using different values for m and B. The results show that the different behaviors of p-Conv (i.e., whether it can perform the task or not) can be modeled mathematically in terms of m and B. The inferred mathematical model reveals interesting patterns able to explain when and why p-Conv succeeds in performing task and, if not, which error pattern it follows.

Prime convolutional model: Breaking the ground for theoretical explainability / Panelli, Francesco; Almhaithawi, Doaa; Cerquitelli, Tania; Bellini, Alessandro. - In: INFORMATION SCIENCES. - ISSN 0020-0255. - ELETTRONICO. - 728:(2025). [10.1016/j.ins.2025.122815]

Prime convolutional model: Breaking the ground for theoretical explainability

Almhaithawi, Doaa;Cerquitelli, Tania;
2025

Abstract

In this paper, we propose a new theoretical approach to Explainable AI. Following the Scientific Method, this approach consists of formulating, on the basis of empirical evidence, a mathematical model to explain and predict the behaviors of Neural Networks. We apply the method to a case study created in a controlled environment, which we call Prime Convolutional Model (p-Conv for short). p-Conv operates on a dataset consisting of the first one million natural numbers and is trained to identify the congruence classes modulo a given integer m. Its architecture uses a convolutional-type neural network that contextually processes a sequence of B consecutive numbers for each input. We take an empirical approach and exploit p-Conv to identify the congruence classes of numbers in a validation set using different values for m and B. The results show that the different behaviors of p-Conv (i.e., whether it can perform the task or not) can be modeled mathematically in terms of m and B. The inferred mathematical model reveals interesting patterns able to explain when and why p-Conv succeeds in performing task and, if not, which error pattern it follows.
File in questo prodotto:
File Dimensione Formato  
INS_D_24_2972_R3.pdf

embargo fino al 22/10/2027

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: Creative commons
Dimensione 915.52 kB
Formato Adobe PDF
915.52 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
Panelli_et_al_IS_2024.pdf

accesso aperto

Tipologia: 1. Preprint / submitted version [pre- review]
Licenza: Pubblico - Tutti i diritti riservati
Dimensione 858.61 kB
Formato Adobe PDF
858.61 kB Adobe PDF Visualizza/Apri
1-s2.0-S002002552500951X-main.pdf

accesso riservato

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 4.08 MB
Formato Adobe PDF
4.08 MB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/3004567