[en] Many of the machine learning tasks rely on centralized learning (CL), which requires the transmission of local datasets from the clients to a parameter server (PS) entailing huge communication overhead. To overcome this, federated learning (FL) has been suggested as a promising tool, wherein the clients send only the model updates to the PS instead of the whole dataset. However, FL demands powerful computational resources from the clients. In practice, not all the clients have sufficient computational resources to participate in training. To address this common scenario, we propose a more efficient approach called hybrid federated and centralized learning (HFCL), wherein only the clients with sufficient resources employ FL, while the remaining ones send their datasets to the PS, which computes the model on behalf of them. Then, the model parameters are aggregated at the PS. To improve the efficiency of dataset transmission, we propose two different techniques: i) increased computation-per-client and ii) sequential data transmission. Notably, the HFCL frameworks outperform FL with up to 20% improvement in the learning accuracy when only half of the clients perform FL while having 50% less communication overhead than CL since all the clients collaborate on the learning process with their datasets.
Disciplines :
Sciences informatiques
Auteur, co-auteur :
Elbir, Ahmet M. ; Duzce University, Department of Electrical and Electronics Engineering, Duzce, Turkey ; University of Luxembourg, SnT, Luxembourg City, Luxembourg
Coleri, Sinem ; Koc University, Department of Electrical and Electronics Engineering, Istanbul, Turkey
Papazafeiropoulos, Anastasios K. ; Duzce University, Department of Electrical and Electronics Engineering, Duzce, Turkey ; University of Hertfordshire, Cis Research Group, Hatfield, United Kingdom
Kourtessis, Pandelis ; University of Hertfordshire, Cis Research Group, Hatfield, United Kingdom
CHATZINOTAS, Symeon ; University of Luxembourg > Interdisciplinary Centre for Security, Reliability and Trust (SNT) > SigCom ; Duzce University, Department of Electrical and Electronics Engineering, Duzce, Turkey
Co-auteurs externes :
yes
Langue du document :
Anglais
Titre :
A Hybrid Architecture for Federated and Centralized Learning
Date de publication/diffusion :
septembre 2022
Titre du périodique :
IEEE Transactions on Cognitive Communications and Networking
eISSN :
2332-7731
Maison d'édition :
Institute of Electrical and Electronics Engineers Inc.
A. M. Elbir, S. Coleri, and K. V. Mishra, "Hybrid federated and centralized learning," in Proc. 29th Eur. Signal Process. Conf. (EUSIPCO), Aug. 2021, pp. 1541-1545.
Y. Lecun, Y. Bengio, and G. Hinton, "Deep learning," Nature, vol. 521, no. 7553, pp. 436-444, 2015.
R. Mayer and H.-A. Jacobsen, "Scalable deep learning on distributed infrastructures: Challenges, techniques, and tools," ACM Comput. Surv., vol. 53, no. 1, pp. 1-37, Feb. 2020.
H. Ye, L. Liang, G. Ye Li, J. Kim, L. Lu, and M. Wu, "Machine learning for vehicular networks: Recent advances and application examples," IEEE Veh. Technol. Mag., vol. 13, no. 2, pp. 94-101, Jun. 2018.
M. S. Mahdavinejad, M. Rezvan, M. Barekatain, P. Adibi, P. Barnaghi, and A. P. Sheth, "Machine learning for Internet of Things data analysis: A survey," Digit. Commun. Netw., vol. 4, no. 3, pp. 161-175, 2018.
A. M. Elbir, B. Soner, and S. Coleri, "Federated learning in vehicular networks," Jun. 2020, arXiv:2006.01412.
A. M. Elbir and K. V. Mishra, "Cognitive learning-aided multiantenna communications," IEEE Wireless Commun., pp. 1-7, May 2022, doi: 10.1109/MWC.008.2100416.
A. M. Elbir and K. V. Mishra, "A survey of deep learning architectures for intelligent reflecting surfaces," Sep. 2020, arXiv:2009.02540.
O. Simeone, "A very brief introduction to machine learning with applications to communication systems," IEEE Trans. Cogn. Commun. Netw., vol. 4, no. 4, pp. 648-664, Dec. 2018.
J. Park, S. Samarakoon, M. Bennis, and M. Debbah, "Wireless network intelligence at the edge," Proc. IEEE, vol. 107, no. 11, pp. 2204-2239, Nov. 2019.
T. Li, A. K. Sahu, A. Talwalkar, and V. Smith, "Federated learning: Challenges, methods, and future directions," IEEE Signal Process. Mag., vol. 37, no. 3, pp. 50-60, May 2020.
M. M. Amiri and D. Gündüz, "Machine learning at the wireless edge: Distributed stochastic gradient descent over-the-air," IEEE Trans. Signal Process., vol. 68, pp. 2155-2169, 2020, doi: 10.1109/TSP.2020.2981904.
M. M. Amiri and D. Gündüz, "Federated learning over wireless fading channels," IEEE Trans. Wireless Commun., vol. 19, no. 5, pp. 3546-3557, May 2020.
A. M. Elbir and K. V. Mishra, "Joint antenna selection and hybrid beamformer design using unquantized and quantized deep learning networks," IEEE Trans. Wireless Commun., vol. 19, no. 3, pp. 1677-1688, Mar. 2020.
H. B. McMahan, E. Moore, D. Ramage, S. Hampson, and B. A. Y. Arcas, "Communication-efficient learning of deep networks from decentralized data," Feb. 2016, arXiv:1602.05629.
Q. Yang, Y. Liu, T. Chen, and Y. Tong, "Federated machine learning: Concept and applications," ACM Trans. Intell. Syst. Technol., vol. 10, no. 2, pp. 1-19, Mar. 2019.
X. Li, K. Huang, W. Yang, S. Wang, and Z. Zhang, "On the convergence of FedAvg on non-IID data," in Proc. Int. Conf. Learn. Represent., 2020, pp. 1-26.
D. Guliani, F. Beaufays, and G. Motta, "Training speech recognition models with federated learning: A quality/cost framework," in Proc. IEEE Int. Conf. Acoust. Speech Signal Process. (ICASSP), Jun. 2021, pp. 3080-3084.
A. M. Elbir and S. Coleri, "Federated learning for hybrid beamforming in mm-wave massive MIMO," IEEE Commun. Lett., vol. 24, no. 12, pp. 2795-2799, Dec. 2020.
A. M. Elbir and S. Coleri, "Federated learning for channel estimation in conventional and RIS-assisted massive MIMO," IEEE Trans. Wireless Commun., early access, Nov. 23, 2021, doi: 10.1109/TWC.2021.3128392.
D. Ma, L. Li, H. Ren, D. Wang, X. Li, and Z. Han, "Distributed rate optimization for intelligent reflecting surface with federated learning," in Proc. IEEE Int. Conf. Commun. Workshops (ICC Workshops), 2020, pp. 1-6.
M. M. Wadu, S. Samarakoon, and M. Bennis, "Federated learning under channel uncertainty: Joint client scheduling and resource allocation," 2020, arXiv:2002.00802.
T. Zeng, O. Semiari, M. Mozaffari, M. Chen, W. Saad, and M. Bennis, "Federated learning in the sky: Joint power allocation and scheduling with UAV swarms," 2020, arXiv:2002.08196.
L. U. Khan, W. Saad, Z. Han, E. Hossain, and C. S. Hong, "Federated learning for Internet of Things: Recent advances, taxonomy, and open challenges," IEEE Commun. Surveys Tuts., vol. 23, no. 3, pp. 1759-1799, 3rd Quart., 2021.
A. M. Elbir, A. K. Papazafeiropoulos, and S. Chatzinotas, "Federated learning for physical layer design," IEEE Commun. Mag., vol. 59, no. 11, pp. 81-87, Nov. 2021.
D. Ye, R. Yu, M. Pan, and Z. Han, "Federated learning in vehicular edge computing: A selective model aggregation approach," IEEE Access, vol. 8, pp. 23920-23935, 2020.
T. Nishio and R. Yonetani, "Client selection for federated learning with heterogeneous resources in mobile edge," in Proc. IEEE Int. Conf. Commun. (ICC), 2019, pp. 1-7.
G. Shi, L. Li, J.Wang,W. Chen, K. Ye, and C. Xu, "HySync: Hybrid federated learning with effective synchronization," in Proc. IEEE 22nd Int. Conf. High Perform. Comput. Commun. IEEE 18th Int. Conf. Smart City IEEE 6th Int. Conf. Data Sci. Syst. (HPCC/SmartCity/DSS), Dec. 2020, pp. 628-633.
X. Zhang, W. Yin, M. Hong, and T. Chen, "Hybrid federated learning: Algorithms and implementation," Dec. 2020, arXiv:2012.12420.
A. Huang et al., "StarFL: Hybrid federated learning architecture for smart urban computing," ACM Trans. Intell. Syst. Technol., vol. 12, no. 4, pp. 1-23, Aug. 2021.
F. Ang, L. Chen, N. Zhao, Y. Chen, W. Wang, and F. R. Yu, "Robust federated learning with noisy communication," IEEE Trans. Commun., vol. 68, no. 6, pp. 3452-3464, Jun. 2020.
F. Haddadpour and M. Mahdavi, "On the convergence of local descent methods in federated learning," Oct. 2019, arXiv:1910.14425.
G. Zhu, Y. Du, D. Gündüz, and K. Huang, "One-bit over-the-air aggregation for communication-efficient federated edge learning: Design and convergence analysis," IEEE Trans. Wireless Commun., vol. 20, no. 3, pp. 2120-2135, Mar. 2021.
C. M. Bishop, "Training with noise is equivalent to Tikhonov regularization," Neural Comput., vol. 7, no. 1, pp. 108-116, Jan. 1995.
X. Wei and C. Shen, "Federated learning over noisy channels," in Proc. IEEE Int. Conf. Commun., Jun. 2021, pp. 1-6.
S. Luo, X. Chen, Q. Wu, Z. Zhou, and S. Yu, "HFEL: Joint edge association and resource allocation for cost-efficient hierarchical federated edge learning," IEEE Trans. Wireless Commun., vol. 19, no. 10, pp. 6535-6548, Oct. 2020.
W. Y. B. Lim et al., "Decentralized edge intelligence: A dynamic resource allocation framework for hierarchical federated learning," IEEE Trans. Parallel Distrib. Syst., vol. 33, no. 3, pp. 536-550, Mar. 2021.
M. Li, T. Zhang, Y. Chen, and A. J. Smola, Efficient Mini-Batch Training for Stochastic Optimization. New York, NY, USA: Assoc. Comput. Mach., Aug. 2014.
S. Zhou and G. Y. Li, "Communication-efficient ADMM-based federated learning," Oct. 2021, arXiv:2110.15318.
S. Sra, S. Nowozin, and S. J. Wright, Optimization for Machine Learning. Cambridge, MA, USA: MIT Press, Sep. 2011.
Y. LeCun, C. Cortes, and C. Burges. "MNIST Handwritten Digit Database." ATT Labs. 2010. [Online]. Available: http://yann.lecun.com/ exdb/mnist
R. Kesten et al., "Lyft Level 5 AV Dataset 2019." 2019. [Online]. Available: https://level5.lyft.com/dataset/ (Accessed: Jun. 1, 2020).
W. Shi, S. Zhou, Z. Niu, M. Jiang, and L. Geng, "Joint device scheduling and resource allocation for latency constrained wireless federated learning," IEEE Trans. Wireless Commun., vol. 20, no. 1, pp. 453-467, Jan. 2021.
T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith, "Federated optimization in heterogeneous networks," in Proc. Int. Conf. Mach. Learn. Syst., vol. 2, 2020, pp. 429-450.
O. Ronneberger, P. Fischer, and T. Brox, "U-Net: Convolutional networks for biomedical image segmentation," in Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Interv., 2015, pp. 234-241.