[en] Multi-objective learning problems occur in all aspects of life and have been studied
for decades, including in the field of machine learning. Many such problems
also exist in distributed settings, where data cannot easily be shared. In recent
years, joint machine learning has been made possible in such settings through the
development of the Federated Learning (FL) paradigm. However, there is as of now
very little research on the general problem of extending the FL concept to multi-
objective learning, limiting such problems to non-cooperative individual learning.
We address this gap by presenting a general framework for multi-objective FL,
based on decomposition (MOFL/D). Our framework addresses the a posteriori
type of multi-objective problem, where user preferences are not known during
the optimisation process, allowing multiple participants to jointly find a set of
solutions, each optimised for some distribution of preferences. We present an
instantiation of the framework and validate it through experiments on a set of
multi-objective benchmarking problems that are extended from well-known single-
objective benchmarks.
Research center :
Interdisciplinary Centre for Security, Reliability and Trust (SnT) > PCOG - Parallel Computing & Optimization Group
Disciplines :
Computer science
Author, co-author :
HARTMANN, Lena Maria ; University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > PCOG
DANOY, Grégoire ; University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Computer Science (DCS)
ALSWAITTI, Mohammed ; University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > PCOG
BOUVRY, Pascal ; University of Luxembourg > Faculty of Science, Technology and Medicine (FSTM) > Department of Computer Science (DCS)
External co-authors :
no
Language :
English
Title :
MOFL/D: A Federated Multi-objective Learning Framework with Decomposition
Publication date :
December 2023
Event name :
International Workshop on Federated Learning in the Age of Foundation Models in Conjunction with NeurIPS 2023 (FL@FM-NeurIPS’23)