Neuro-symbolic techniques such as logic tensor networks (LTNs) enable the integration of symbolic knowledge to improve the learning capabilities of deep neural networks. LTNs in particular ground first-order logic languages into differentiable tensor operations, redefining learning as maximizing the satisfiability of a grounded theory. Despite the promising results achieved so far, the optimization task is highly sensitive to the choice of functions for grounding logical operators and aggregators, limiting their practical adoption. The present study focuses on learning in the presence of class imbalance (in object detection tasks, class imbalance arises between background vs foreground samples). In particular, we seek to combine the recently proposed logLTN with the weighting scheme introduced by the focal loss as an enhancement of the original cross-entropy loss. Preliminary experiments on an object detection benchmark show that the focal logLTN aggregator achieves higher performance and stability than its standard counterpart, with potential application in many other practical scenarios.

Enhancing Neuro-Symbolic Integration with Focal Loss: A Study on Logic Tensor Networks / Piano, Luca; Manigrasso, Francesco; Russo, Alessandro; Morra, Lia. - STAMPA. - 14980:(2024), pp. 14-23. (Intervento presentato al convegno 18th International Conference on Neuro-symbolic Learning and Reasoning tenutosi a Barcelona (ESP) nel September 9–12, 2024) [10.1007/978-3-031-71170-1_2].

Enhancing Neuro-Symbolic Integration with Focal Loss: A Study on Logic Tensor Networks

Luca Piano;Francesco Manigrasso;Alessandro Russo;Lia Morra
2024

Abstract

Neuro-symbolic techniques such as logic tensor networks (LTNs) enable the integration of symbolic knowledge to improve the learning capabilities of deep neural networks. LTNs in particular ground first-order logic languages into differentiable tensor operations, redefining learning as maximizing the satisfiability of a grounded theory. Despite the promising results achieved so far, the optimization task is highly sensitive to the choice of functions for grounding logical operators and aggregators, limiting their practical adoption. The present study focuses on learning in the presence of class imbalance (in object detection tasks, class imbalance arises between background vs foreground samples). In particular, we seek to combine the recently proposed logLTN with the weighting scheme introduced by the focal loss as an enhancement of the original cross-entropy loss. Preliminary experiments on an object detection benchmark show that the focal logLTN aggregator achieves higher performance and stability than its standard counterpart, with potential application in many other practical scenarios.
2024
978-3-031-71169-5
978-3-031-71170-1
File in questo prodotto:
File Dimensione Formato  
FocalLTN___NeSy24.pdf

embargo fino al 10/09/2025

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: PUBBLICO - Tutti i diritti riservati
Dimensione 440.4 kB
Formato Adobe PDF
440.4 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
978-3-031-71170-1_2.pdf

non disponibili

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 331.19 kB
Formato Adobe PDF
331.19 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2990033