Neuro-symbolic techniques such as logic tensor networks (LTNs) enable the integration of symbolic knowledge to improve the learning capabilities of deep neural networks. LTNs in particular ground first-order logic languages into differentiable tensor operations, redefining learning as maximizing the satisfiability of a grounded theory. Despite the promising results achieved so far, the optimization task is highly sensitive to the choice of functions for grounding logical operators and aggregators, limiting their practical adoption. The present study focuses on learning in the presence of class imbalance (in object detection tasks, class imbalance arises between background vs foreground samples). In particular, we seek to combine the recently proposed logLTN with the weighting scheme introduced by the focal loss as an enhancement of the original cross-entropy loss. Preliminary experiments on an object detection benchmark show that the focal logLTN aggregator achieves higher performance and stability than its standard counterpart, with potential application in many other practical scenarios.
Enhancing Neuro-Symbolic Integration with Focal Loss: A Study on Logic Tensor Networks / Piano, Luca; Manigrasso, Francesco; Russo, Alessandro; Morra, Lia. - STAMPA. - 14980:(2024), pp. 14-23. (Intervento presentato al convegno 18th International Conference on Neuro-symbolic Learning and Reasoning tenutosi a Barcelona (ESP) nel September 9–12, 2024) [10.1007/978-3-031-71170-1_2].
Enhancing Neuro-Symbolic Integration with Focal Loss: A Study on Logic Tensor Networks
Luca Piano;Francesco Manigrasso;Alessandro Russo;Lia Morra
2024
Abstract
Neuro-symbolic techniques such as logic tensor networks (LTNs) enable the integration of symbolic knowledge to improve the learning capabilities of deep neural networks. LTNs in particular ground first-order logic languages into differentiable tensor operations, redefining learning as maximizing the satisfiability of a grounded theory. Despite the promising results achieved so far, the optimization task is highly sensitive to the choice of functions for grounding logical operators and aggregators, limiting their practical adoption. The present study focuses on learning in the presence of class imbalance (in object detection tasks, class imbalance arises between background vs foreground samples). In particular, we seek to combine the recently proposed logLTN with the weighting scheme introduced by the focal loss as an enhancement of the original cross-entropy loss. Preliminary experiments on an object detection benchmark show that the focal logLTN aggregator achieves higher performance and stability than its standard counterpart, with potential application in many other practical scenarios.File | Dimensione | Formato | |
---|---|---|---|
FocalLTN___NeSy24.pdf
embargo fino al 10/09/2025
Tipologia:
2. Post-print / Author's Accepted Manuscript
Licenza:
PUBBLICO - Tutti i diritti riservati
Dimensione
440.4 kB
Formato
Adobe PDF
|
440.4 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
978-3-031-71170-1_2.pdf
non disponibili
Tipologia:
2a Post-print versione editoriale / Version of Record
Licenza:
Non Pubblico - Accesso privato/ristretto
Dimensione
331.19 kB
Formato
Adobe PDF
|
331.19 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.
https://hdl.handle.net/11583/2990033