To increase the trustworthiness of deep neural networks, it is critical to improve the understanding of how they make decisions. This paper introduces a novel unsupervised concept-based model for image classification, named Learnable Concept-Based Model (LCBM) which models concepts as random variables within a Bernoulli latent space. Unlike traditional methods that either require extensive human supervision or suffer from limited scalability, our approach employs a reduced number of concepts without sacrificing performance. We demonstrate that LCBM surpasses existing unsupervised concept-based models in generalization capability and nearly matches the performance of black-box models. The proposed concept representation enhances information retention and aligns more closely with human understanding. A user study demonstrates the discovered concepts are also more intuitive for humans to interpret. Finally, despite the use of concept embeddings, we maintain model interpretability by means of a local linear combination of concepts.
Towards Better Generalization and Interpretability in Unsupervised Concept-Based Models / De Santis, Francesco; Bich, Philippe; Ciravegna, Gabriele; Barbiero, Pietro; Cerquitelli &, Tania; Giordano, Danilo. - 16015:(2026), pp. 478-494. ( European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML PKDD) Porto (PRT) September 15–19, 2025) [10.1007/978-3-032-06066-2_28].
Towards Better Generalization and Interpretability in Unsupervised Concept-Based Models
Francesco De Santis;Philippe Bich;Gabriele Ciravegna;Danilo Giordano
2026
Abstract
To increase the trustworthiness of deep neural networks, it is critical to improve the understanding of how they make decisions. This paper introduces a novel unsupervised concept-based model for image classification, named Learnable Concept-Based Model (LCBM) which models concepts as random variables within a Bernoulli latent space. Unlike traditional methods that either require extensive human supervision or suffer from limited scalability, our approach employs a reduced number of concepts without sacrificing performance. We demonstrate that LCBM surpasses existing unsupervised concept-based models in generalization capability and nearly matches the performance of black-box models. The proposed concept representation enhances information retention and aligns more closely with human understanding. A user study demonstrates the discovered concepts are also more intuitive for humans to interpret. Finally, despite the use of concept embeddings, we maintain model interpretability by means of a local linear combination of concepts.| File | Dimensione | Formato | |
|---|---|---|---|
|
978-3-032-06066-2_28.pdf
accesso riservato
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
7.03 MB
Formato
Adobe PDF
|
7.03 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
|
_ECML25__Learnable_Concept_Based_Model.pdf
embargo fino al 04/10/2026
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
Pubblico - Tutti i diritti riservati
Dimensione
1.7 MB
Formato
Adobe PDF
|
1.7 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/3005352
