skip to main content
research-article

Inexact-ADMM Based Federated Meta-Learning for Fast and Continual Edge Learning

Published: 26 July 2021 Publication History

Abstract

In order to meet the requirements for performance, safety, and latency in many IoT applications, intelligent decisions must be made right here right now at the network edge. However, the constrained resources and limited local data amount pose significant challenges to the development of edge AI. To overcome these challenges, we explore continual edge learning capable of leveraging the knowledge transfer from previous tasks. Aiming to achieve fast and continual edge learning, we propose a platform-aided federated meta-learning architecture where edge nodes collaboratively learn a meta-model, aided by the knowledge transfer from prior tasks. The edge learning problem is cast as a regularized optimization problem, where the valuable knowledge learned from previous tasks is extracted as regularization. Then, we devise an ADMM based federated meta-learning algorithm, namely ADMM-FedMeta, where ADMM offers a natural mechanism to decompose the original problem into many subproblems which can be solved in parallel across edge nodes and the platform. Further, a variant of inexact-ADMM method is employed where the subproblems are 'solved' via linear approximation as well as Hessian estimation to reduce the computational cost per round to O(n). We provide a comprehensive analysis of ADMM-FedMeta, in terms of the convergence properties, the rapid adaptation performance, and the forgetting effect of prior knowledge transfer, for the general non-convex case. Extensive experimental studies demonstrate the effectiveness and efficiency of ADMM-FedMeta, and showcase that it substantially outperforms the existing baselines.

References

[1]
Rina Foygel Barber and Emil Y Sidky. 2020. Convergence for nonconvex ADMM, with applications to CT imaging. arXiv preprint arXiv:2006.07278 (2020).
[2]
Pablo Barros, German I Parisi, Di Fu, Xun Liu, and Stefan Wermter. 2018. Expectation learning for adaptive crossmodal stimuli association. arXiv preprint arXiv:1801.07654 (2018).
[3]
Stephen Boyd, Neal Parikh, and Eric Chu. 2011. Distributed optimization and statistical learning via the alternating direction method of multipliers. Now Publishers Inc.
[4]
Lev M Bregman. 1967. The relaxation method of finding the common point of convex sets and its application to the solution of problems in convex programming. USSR computational mathematics and mathematical physics 7, 3 (1967), 200--217.
[5]
Fei Chen, Mi Luo, Zhenhua Dong, Zhenguo Li, and Xiuqiang He. 2018. Federated meta-learning with fast convergence and efficient communication. arXiv preprint arXiv:1802.07876 (2018).
[6]
Liam Collins, Aryan Mokhtari, and Sanjay Shakkottai. 2020. Distribution-Agnostic Model-Agnostic Meta-Learning. arXiv preprint arXiv:2002.04766 (2020).
[7]
Alireza Fallah, Aryan Mokhtari, and Asuman Ozdaglar. 2020. On the convergence theory of gradient-based model-agnostic meta-learning algorithms. In International Conference on Artificial Intelligence and Statistics. 1082--1092.
[8]
Alireza Fallah, Aryan Mokhtari, and Asuman Ozdaglar. 2020. Personalized federated learning: A meta-learning approach. arXiv preprint arXiv:2002.07948 (2020).
[9]
Chelsea Finn, Pieter Abbeel, and Sergey Levine. 2017. Model-agnostic meta-learning for fast adaptation of deep networks. arXiv preprint arXiv:1703.03400 (2017).
[10]
Chelsea Finn, Kelvin Xu, and Sergey Levine. 2018. Probabilistic model-agnostic meta-learning. In Advances in Neural Information Processing Systems. 9516--9527.
[11]
Robert M French. 1999. Catastrophic forgetting in connectionist networks. Trends in cognitive sciences 3, 4 (1999), 128--135.
[12]
Mingyi Hong, Zhi-Quan Luo, and Meisam Razaviyayn. 2016. Convergence analysis of alternating direction method of multipliers for a family of nonconvex problems. SIAM Journal on Optimization 26, 1 (2016), 337--364.
[13]
Kaiyi Ji, Junjie Yang, and Yingbin Liang. 2020. Multi-Step Model-Agnostic Meta-Learning: Convergence and Improved Algorithms. arXiv preprint arXiv:2002.07836 (2020).
[14]
Bo Jiang, Tianyi Lin, Shiqian Ma, and Shuzhong Zhang. 2019. Structured nonconvex and nonsmooth optimization: algorithms and iteration complexity analysis. Computational Optimization and Applications 72, 1 (2019), 115--157.
[15]
Yihan Jiang, Jakub Konečny, Keith Rush, and Sreeram Kannan. 2019. Improving federated learning personalization via model agnostic meta learning. arXiv preprint arXiv:1909.12488 (2019).
[16]
Heechul Jung, Jeongwoo Ju, Minju Jung, and Junmo Kim. 2016. Less-forgetting learning in deep neural networks. arXiv preprint arXiv:1607.00122 (2016).
[17]
James Kirkpatrick, Razvan Pascanu, Neil Rabinowitz, Joel Veness, Guillaume Desjardins, Andrei A Rusu, Kieran Milan, John Quan, Tiago Ramalho, Agnieszka Grabska-Barwinska, et al. 2017. Overcoming catastrophic forgetting in neural networks. Proceedings of the national academy of sciences 114, 13 (2017), 3521--3526.
[18]
R Krishnan and Prasanna Balaprakash. 2020. Meta Continual Learning via Dynamic Programming. arXiv preprint arXiv:2008.02219 (2020).
[19]
Alex Krizhevsky, Geoffrey Hinton, et al. 2009. Learning multiple layers of features from tiny images. (2009).
[20]
Alessandro Lanza, Serena Morigi, Ivan Selesnick, and Fiorella Sgallari. 2017. Nonconvex nonsmooth optimization via convex-nonconvex majorization-minimization. Numer. Math. 136, 2 (2017), 343--381.
[21]
Sen Lin, Guang Yang, and Junshan Zhang. 2020. A Collaborative Learning Framework via Federated Meta-Learning. arXiv preprint arXiv:2001.03229 (2020).
[22]
David Lopez-Paz and Marc'Aurelio Ranzato. 2017. Gradient episodic memory for continual learning. In Advances in neural information processing systems. 6467--6476.
[23]
Sindri Magnússon, Pradeep Chathuranga Weeraddana, Michael G Rabbat, and Carlo Fischione. 2015. On the convergence of alternating direction lagrangian methods for nonconvex structured optimization problems. IEEE Transactions on Control of Network Systems 3, 3 (2015), 296--309.
[24]
Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise Aguera y Arcas. 2017. Communication-efficient learning of deep networks from decentralized data. In Artificial Intelligence and Statistics. 1273--1282.
[25]
Mahesh Chandra Mukkamala, Peter Ochs, Thomas Pock, and Shoham Sabach. 2020. Convex-Concave Backtracking for Inertial Bregman Proximal Gradient Algorithms in Nonconvex Optimization. SIAM Journal on Mathematics of Data Science 2, 3 (2020), 658--682.
[26]
Alex Nichol, Joshua Achiam, and John Schulman. 2018. On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018).
[27]
German I Parisi, Ronald Kemker, Jose L Part, Christopher Kanan, and Stefan Wermter. 2019. Continual lifelong learning with neural networks: A review. Neural Networks 113 (2019), 54--71.
[28]
Jihong Park, Sumudu Samarakoon, Mehdi Bennis, and Mérouane Debbah. 2019. Wireless network intelligence at the edge. Proc. IEEE 107, 11 (2019), 2204--2239.
[29]
Aniruddh Raghu, Maithra Raghu, Samy Bengio, and Oriol Vinyals. 2019. Rapid learning or feature reuse? towards understanding the effectiveness of maml. arXiv preprint arXiv:1909.09157 (2019).
[30]
Sachin Ravi and Hugo Larochelle. 2016. Optimization as a model for few-shot learning. (2016).
[31]
Jonathan Schwarz, Jelena Luketina, Wojciech M Czarnecki, Agnieszka Grabska-Barwinska, Yee Whye Teh, Razvan Pascanu, and Raia Hadsell. 2018. Progress & compress: A scalable framework for continual learning. arXiv preprint arXiv:1805.06370 (2018).
[32]
Si Si, Dacheng Tao, and Bo Geng. 2009. Bregman divergence-based regularization for transfer subspace learning. IEEE Transactions on Knowledge and Data Engineering 22, 7 (2009), 929--942.
[33]
Xingyou Song, Wenbo Gao, Yuxiang Yang, Krzysztof Choromanski, Aldo Pacchiano, and Yunhao Tang. 2019. Es-maml: Simple hessian-free meta learning. arXiv preprint arXiv:1910.01215 (2019).
[34]
Masaaki Takada and Hironori Fujisawa. 2020. Transfer Learning via ℓ1 Regularization. Advances in Neural Information Processing Systems 33 (2020).
[35]
Fenghui Wang, Wenfei Cao, and Zongben Xu. 2018. Convergence of multi-block Bregman ADMM for nonconvex composite problems. Science China Information Sciences 61, 12 (2018), 122101.
[36]
Fenghui Wang, Zongben Xu, and Hong-Kun Xu. 2014. Convergence of Bregman alternating direction method with multipliers for nonconvex composite problems. arXiv preprint arXiv:1410.8625 (2014).
[37]
Junxiang Wang, Fuxun Yu, Xiang Chen, and Liang Zhao. 2019. Admm for efficient deep learning with global convergence. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining. 111--119.
[38]
Lingxiao Wang, Qi Cai, Zhuoran Yang, and Zhaoran Wang. 2020. On the Global Optimality of Model-Agnostic Meta-Learning. arXiv preprint arXiv:2006.13182 (2020).
[39]
Yu Wang, Wotao Yin, and Jinshan Zeng. 2019. Global convergence of ADMM in nonconvex nonsmooth optimization. Journal of Scientific Computing 78, 1 (2019), 29--63.
[40]
Qiong Wu, Kaiwen He, and Xu Chen. 2020. Personalized federated learning for intelligent iot applications: A cloud-edge based framework. IEEE Computer Graphics and Applications (2020).
[41]
Han Xiao, Kashif Rasul, and Roland Vollgraf. 2017. Fashion-mnist: a novel image dataset for benchmarking machine learning algorithms. arXiv preprint arXiv:1708.07747 (2017).
[42]
Dong Yu, Kaisheng Yao, Hang Su, Gang Li, and Frank Seide. 2013. KL-divergence regularized deep neural network adaptation for improved large vocabulary speech recognition. In 2013 IEEE International Conference on Acoustics, Speech and Signal Processing. IEEE, 7893--7897.
[43]
Sheng Yue, Ju Ren, Jiang Xin, Sen Lin, and Junshan Zhang. 2020. Inexact-ADMM based Federated Meta-Learning for Fast and Continual Edge Learning. arXiv:2012.08677 [cs.LG]
[44]
Friedemann Zenke, Ben Poole, and Surya Ganguli. 2017. Continual learning through synaptic intelligence. Proceedings of machine learning research 70 (2017), 3987.
[45]
Xinwei Zhang, Mingyi Hong, Sairaj Dhople, Wotao Yin, and Yang Liu. 2020. FedPD: A Federated Learning Framework with Optimal Rates and Adaptivity to Non-IID Data. arXiv preprint arXiv:2005.11418 (2020).
[46]
Zhaofeng Zhang, Sen Lin, Mehmet Dedeoglu, Kemi Ding, and Junshan Zhang. 2020. Data-driven Distributionally Robust Optimization for Edge Intelligence. In IEEE INFOCOM 2020-IEEE Conference on Computer Communications. IEEE, 2619--2628.
[47]
Wenbo Zheng, Lan Yan, Chao Gou, and Fei-Yue Wang. 2020. Federated Meta-Learning for Fraudulent Credit Card Detection. Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence (IJCAI-20) (2020).
[48]
Zhi Zhou, Xu Chen, En Li, Liekang Zeng, Ke Luo, and Junshan Zhang. 2019. Edge intelligence: Paving the last mile of artificial intelligence with edge computing. Proc. IEEE 107, 8 (2019), 1738--1762.

Cited By

View all
  • (2024)Space–Air–Ground–Sea Integrated Network with Federated LearningRemote Sensing10.3390/rs1609164016:9(1640)Online publication date: 4-May-2024
  • (2024)Symmetric ADMM-Based Federated Learning with a Relaxed StepMathematics10.3390/math1217266112:17(2661)Online publication date: 27-Aug-2024
  • (2024)Resource-Efficient Heterogenous Federated Continual Learning on Edge2024 Design, Automation & Test in Europe Conference & Exhibition (DATE)10.23919/DATE58400.2024.10546889(1-6)Online publication date: 25-Mar-2024
  • Show More Cited By

Index Terms

  1. Inexact-ADMM Based Federated Meta-Learning for Fast and Continual Edge Learning

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      MobiHoc '21: Proceedings of the Twenty-second International Symposium on Theory, Algorithmic Foundations, and Protocol Design for Mobile Networks and Mobile Computing
      July 2021
      286 pages
      ISBN:9781450385589
      DOI:10.1145/3466772
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 26 July 2021

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. ADMM
      2. continual learning
      3. edge intelligence
      4. federated meta-learning
      5. regularization

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Funding Sources

      • Natural Science Foundation of Hunan Province, China
      • Young Elite Scientists Sponsorship Program by CAST
      • 111 Project
      • National Key R&D Program of China
      • National Natural Science Foundation of China
      • Young Talents Plan of Hunan Province of China

      Conference

      MobiHoc '21
      Sponsor:

      Acceptance Rates

      MobiHoc '21 Paper Acceptance Rate 28 of 139 submissions, 20%;
      Overall Acceptance Rate 296 of 1,843 submissions, 16%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)103
      • Downloads (Last 6 weeks)3
      Reflects downloads up to 19 Oct 2024

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Space–Air–Ground–Sea Integrated Network with Federated LearningRemote Sensing10.3390/rs1609164016:9(1640)Online publication date: 4-May-2024
      • (2024)Symmetric ADMM-Based Federated Learning with a Relaxed StepMathematics10.3390/math1217266112:17(2661)Online publication date: 27-Aug-2024
      • (2024)Resource-Efficient Heterogenous Federated Continual Learning on Edge2024 Design, Automation & Test in Europe Conference & Exhibition (DATE)10.23919/DATE58400.2024.10546889(1-6)Online publication date: 25-Mar-2024
      • (2024)Federated Deep Equilibrium Learning: Harnessing Compact Global Representations to Enhance PersonalizationProceedings of the 33rd ACM International Conference on Information and Knowledge Management10.1145/3627673.3679752(1132-1142)Online publication date: 21-Oct-2024
      • (2024)Game-Theoretic Federated Meta-learning for Blockchain-Assisted Metaverse2024 IEEE Wireless Communications and Networking Conference (WCNC)10.1109/WCNC57260.2024.10571133(1-6)Online publication date: 21-Apr-2024
      • (2024)Federated PCA on Grassmann Manifold for IoT Anomaly DetectionIEEE/ACM Transactions on Networking10.1109/TNET.2024.342378032:5(4456-4471)Online publication date: Oct-2024
      • (2024)Towards Resource-Efficient Edge AI: From Federated Learning to Semi-Supervised Model PersonalizationIEEE Transactions on Mobile Computing10.1109/TMC.2023.331618923:5(6104-6115)Online publication date: May-2024
      • (2024)Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM2024 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML)10.1109/SaTML59370.2024.00029(443-471)Online publication date: 9-Apr-2024
      • (2024)A Blockchain-Based Reliable Federated Meta-Learning for Metaverse: A Dual Game FrameworkIEEE Internet of Things Journal10.1109/JIOT.2024.338309611:12(22697-22715)Online publication date: 15-Jun-2024
      • (2024)Efficient knowledge management for heterogeneous federated continual learning on resource-constrained edge devicesFuture Generation Computer Systems10.1016/j.future.2024.02.018156(16-29)Online publication date: Jul-2024
      • Show More Cited By

      View Options

      Get Access

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media