Convergence bounds are one of the main tools to obtain information on the performance of a distributed machine learning task, before running the task itself. In this work, we perform a set of experiments to assess to which extent, and in which way, such bounds can predict and improve the performance of real-world distributed (namely, federated) learning tasks. We find that, as can be expected given the way they are obtained, bounds are quite loose and their relative magnitude reflects the training rather than the testing loss. More unexpectedly, we find that some of the quantities appearing in the bounds turn out to be very useful to identify the clients that are most likely to contribute to the learning process, without requiring the disclosure of any information about the quality or size of their datasets. This suggests that further research is warranted on the ways – often counter-intuitive – in which convergence bounds can be exploited to improve the performance of real-world distributed learning tasks.

Unexpectedly Useful: Convergence Bounds And Real-World Distributed Learning / Malandrino, Francesco; Chiasserini, Carla Fabiana. - STAMPA. - (2023), pp. 76-79. (Intervento presentato al convegno 2023 15th International Conference on Machine Learning and Computing (ICMLC 2023) tenutosi a Zhuhai (China) nel Feb. 2023) [10.1145/3587716.3587728].

Unexpectedly Useful: Convergence Bounds And Real-World Distributed Learning

Carla Fabiana Chiasserini
2023

Abstract

Convergence bounds are one of the main tools to obtain information on the performance of a distributed machine learning task, before running the task itself. In this work, we perform a set of experiments to assess to which extent, and in which way, such bounds can predict and improve the performance of real-world distributed (namely, federated) learning tasks. We find that, as can be expected given the way they are obtained, bounds are quite loose and their relative magnitude reflects the training rather than the testing loss. More unexpectedly, we find that some of the quantities appearing in the bounds turn out to be very useful to identify the clients that are most likely to contribute to the learning process, without requiring the disclosure of any information about the quality or size of their datasets. This suggests that further research is warranted on the ways – often counter-intuitive – in which convergence bounds can be exploited to improve the performance of real-world distributed learning tasks.
File in questo prodotto:
File Dimensione Formato  
ICMLC_bounds-3.pdf

accesso aperto

Tipologia: 2. Post-print / Author's Accepted Manuscript
Licenza: PUBBLICO - Tutti i diritti riservati
Dimensione 781.61 kB
Formato Adobe PDF
781.61 kB Adobe PDF Visualizza/Apri
Chiasserini-Unexpectedly.pdf

non disponibili

Tipologia: 2a Post-print versione editoriale / Version of Record
Licenza: Non Pubblico - Accesso privato/ristretto
Dimensione 807.5 kB
Formato Adobe PDF
807.5 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11583/2973549