FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning
Abstract
:1. Introduction
- FedOpt utilises the novel Sparse Compression Algorithm (SCA) in order to reduce the communication overhead. In particular, SCA extends the existing top-k gradient compression technique and enables downstream compression with a novel mechanism.
- FedOpt adopts a lightweight homomorphic encryption for efficient and secure aggregation of the gradients. In particular, FedOpt provides a concrete abstract, where additively homomorphic encryption is completely utilised in order to eliminate the key-switching operation and to increase the space for plain-text.
- To further ensure the privacy of local users from the collusion of adversaries, FedOpt uses a differential-privacy scheme based on Laplace mechanism in order to keep the originality of local gradients.
- FedOpt tolerates user drops during the training process with negligible amounts of accuracy losses. Furthermore, the performance evaluation demonstrates the training accuracy of FedOpt in real-life scenarios as well as its efficient communication and low computation overhead.
2. System Model and Problem Statement
2.1. System Model
2.2. Problem Statement
3. Preliminaries
3.1. Federated Learning
- The learned model is shared between the users and the cloud server. However, the training data which is distributed on each user is not available to the cloud server.
- Instead of the cloud server, the training of learning model occurs on each user. The cloud server receives the local gradients and aggregates these gradients to obtain a global gradient and then send this global gradient back to all the users.
3.2. Additively Homomorphic Encryption
3.3. Differential Privacy
3.4. Laplace Mechanism
4. Federated Optimisation (FedOpt)
4.1. Sparse Compression Algorithm (SCA)
SCA Technique
Algorithm 1: SCA: Communication Efficiency in FedOpt |
Input: temporal vector , Sparsity Fraction Output: sparse temporal 1 Initialisation: 2← (); ← () 3← mean (); ← mean () 4 if≥ 5 then 6 return (
← ( ≥ min()) 7 else return (← ( ≤ min())) 8 end |
4.2. Gradient Aggregation in FedOpt
Algorithm 2: Pseudocode of Privacy Preserving |
4.3. Efficiency and Privacy in FedOpt
4.3.1. Initialisation Phase
4.3.2. Encryption Phase
4.3.3. Aggregation Phase
4.3.4. Decryption Phase
Algorithm 3: FedOpt: Communication-Efficiency and Privacy-Preserving |
5. FedOpt Evaluation
5.1. Accuracy Test
5.2. Communication Efficiency
5.3. Analysis of Communication Efficiency w.r.t Accuracy
5.4. Computation Overhead
6. Related Work and Discussions
Functional Comparison
7. Conclusions
Author Contributions
Funding
Conflicts of Interest
Abbreviations
FedOpt | Federated Optimisation |
AI | Artificial Intelligence |
FL | Federated Learning |
DL | Deep Learning |
SGD | Stochastic Gradient Descent |
DNN | Deep Neural Network |
SCA | Sparse Compression Algorithm |
- | Privacy Budget on Differential Privacy |
⨿ | users |
Secret Key | |
Expected Loss | |
Initial Parameters | |
Global Parameters | |
Parameter Vector |
References
- Müller, V.C.; Bostrom, N. Future progress in artificial intelligence: A survey of expert opinion. In Fundamental Issues of Artificial Intelligence; Springer: Cham, Switzerland, 2016; pp. 555–572. [Google Scholar]
- Cheng, E.J.; Chou, K.P.; Rajora, S.; Jin, B.H.; Tanveer, M.; Lin, C.T.; Young, K.Y.; Lin, W.C.; Prasad, M. Deep sparse representation classifier for facial recognition and detection system. Pattern Recognit. Lett. 2019, 125, 71–77. [Google Scholar] [CrossRef]
- Li, D.; Zhao, D.; Zhang, Q.; Chen, Y. Reinforcement learning and deep learning based lateral control for autonomous driving [application notes]. IEEE Comput. Intell. Mag. 2019, 14, 83–98. [Google Scholar] [CrossRef]
- Yang, Y.; Li, W.; Gulliver, T.A.; Li, S. Bayesian Deep Learning Based Probabilistic Load Forecasting in Smart Grids. IEEE Trans. Ind. Inform. 2019, 16, 4703–4713. [Google Scholar] [CrossRef]
- Kumar, D.S.; Thilagam, P.S. Approaches and challenges of privacy preserving search over encrypted data. Inf. Syst. 2019, 81, 63–81. [Google Scholar] [CrossRef]
- McMahan, B.; Ramage, D. Federated learning: Collaborative machine learning without centralized training data. Google Res. Blog 2017, 3. [Google Scholar]
- Dean, J.; Corrado, G.; Monga, R.; Chen, K.; Devin, M.; Mao, M.; Ranzato, M.; Senior, A.; Tucker, P.; Le, Q.V.; et al. Large scale distributed deep networks. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2012; pp. 1223–1231. [Google Scholar]
- Sattler, F.; Wiedemann, S.; Müller, K.R.; Samek, W. Robust and communication-efficient federated learning from non-iid data. IEEE Trans. Neural Netw. Learn. Syst. 2019. [Google Scholar] [CrossRef] [Green Version]
- Bhagoji, A.N.; Chakraborty, S.; Mittal, P.; Calo, S. Analyzing federated learning through an adversarial lens. arXiv 2018, arXiv:1811.12470. [Google Scholar]
- Nasr, M.; Shokri, R.; Houmansadr, A. Comprehensive privacy analysis of deep learning: Stand-alone and federated learning under passive and active white-box inference attacks. arXiv 2018, arXiv:1812.00910. [Google Scholar]
- Hao, M.; Li, H.; Luo, X.; Xu, G.; Yang, H.; Liu, S. Efficient and privacy-enhanced federated learning for industrial artificial intelligence. IEEE Trans. Ind. Inform. 2019. [Google Scholar] [CrossRef]
- Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; d’Oliveira, R.G.; et al. Advances and open problems in federated learning. arXiv 2019, arXiv:1912.04977. [Google Scholar]
- Zhang, X.; Chen, X.; Liu, J.; Xiang, Y. DeepPAR and DeepDPA: Privacy-Preserving and Asynchronous Deep Learning for Industrial IoT. IEEE Trans. Ind. Inform. 2019, 16, 2081–2090. [Google Scholar] [CrossRef]
- Tripathy, A.; Wang, Y.; Ishwar, P. Privacy-preserving adversarial networks. In Proceedings of the 2019 57th Annual Allerton Conference on Communication, Control, and Computing (Allerton), Monticello, IL, USA, 24–27 September 2019; pp. 495–505. [Google Scholar]
- Dwork, C.; McSherry, F.; Nissim, K.; Smith, A. Calibrating noise to sensitivity in private data analysis. J. Priv. Confid. 2016, 7, 17–51. [Google Scholar]
- Aji, A.F.; Heafield, K. Sparse communication for distributed gradient descent. arXiv 2017, arXiv:1704.05021. [Google Scholar]
- Lin, Y.; Han, S.; Mao, H.; Wang, Y.; Dally, W.J. Deep gradient compression: Reducing the communication bandwidth for distributed training. arXiv 2017, arXiv:1712.01887. [Google Scholar]
- Cheng, Y.; Wang, D.; Zhou, P.; Zhang, T. A survey of model compression and acceleration for deep neural networks. arXiv 2017, arXiv:1710.09282. [Google Scholar]
- Shokri, R.; Shmatikov, V. Privacy-preserving deep learning. In Proceedings of the 22nd ACM SIGSAC Conference on Computer and Communications Security, Denver, CO, USA, 12–16 October 2015; pp. 1310–1321. [Google Scholar]
- Peralta, G.; Cid-Fuentes, R.G.; Bilbao, J.; Crespo, P.M. Homomorphic Encryption and Network Coding in IoT Architectures: Advantages and Future Challenges. Electronics 2019, 8, 827. [Google Scholar] [CrossRef] [Green Version]
- Vadhan, S. The complexity of differential privacy. In Tutorials on the Foundations of Cryptography; Springer: Cham, Switzerland, 2017; pp. 347–450. [Google Scholar]
- Roy, P. A homomorphism based zero knowledge proof of authentication for chinese remainder theorem based secret sharing. In Annual Convention of the Computer Society of India; Springer: Singapore, 2018; pp. 562–572. [Google Scholar]
- McMahan, H.B.; Moore, E.; Ramage, D.; Hampson, S. Communication-efficient learning of deep networks from decentralized data. arXiv 2016, arXiv:1602.05629. [Google Scholar]
- Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-preserving deep learning via additively homomorphic encryption. IEEE Trans. Inf. Forensics Secur. 2017, 13, 1333–1345. [Google Scholar]
- Fazio, N.; Gennaro, R.; Jafarikhah, T.; Skeith, W.E. Homomorphic secret sharing from paillier encryption. In International Conference on Provable Security; Springer: Cham, Switzerland, 2017; pp. 381–399. [Google Scholar]
- Chilimbi, T.; Suzue, Y.; Apacible, J.; Kalyanaraman, K. Project adam: Building an efficient and scalable deep learning training system. In Proceedings of the 11th USENIX Symposium on Operating Systems Design and Implementation (OSDI 14), Broomfield, CO, USA, 6–8 October 2014; pp. 571–582. [Google Scholar]
- Zinkevich, M.; Weimer, M.; Li, L.; Smola, A.J. Parallelized stochastic gradient descent. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2010; pp. 2595–2603. [Google Scholar]
- Li, M.; Andersen, D.G.; Smola, A.J.; Yu, K. Communication efficient distributed machine learning with the parameter server. In Advances in Neural Information Processing Systems; The MIT Press: Cambridge, MA, USA, 2014; pp. 19–27. [Google Scholar]
- Aono, Y.; Hayashi, T.; Wang, L.; Moriai, S. Privacy-preserving deep learning: Revisited and enhanced. In International Conference on Applications and Techniques in Information Security; Springer: Singapore, 2017; pp. 100–110. [Google Scholar]
- Agarwal, N.; Suresh, A.T.; Yu, F.X.X.; Kumar, S.; McMahan, B. cpSGD: Communication-efficient and differentially-private distributed SGD. In Advances in Neural Information Processing Systems; Neural Information Processing Systems Conference: Montréal, QC, Canada, 2018; pp. 7564–7575. [Google Scholar]
- Dörner, S.; Cammerer, S.; Hoydis, J.; ten Brink, S. Deep learning based communication over the air. IEEE J. Sel. Top. Signal Process. 2017, 12, 132–143. [Google Scholar] [CrossRef] [Green Version]
- Nishio, T.; Yonetani, R. Client selection for federated learning with heterogeneous resources in mobile edge. In Proceedings of the 2019 IEEE International Conference on Communications (ICC 2019), Shanghai, China, 20–24 May 2019; pp. 1–7. [Google Scholar]
- Konečný, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
- Bernstein, J.; Wang, Y.X.; Azizzadenesheli, K.; Anandkumar, A. signSGD: Compressed optimisation for non-convex problems. arXiv 2018, arXiv:1802.04434. [Google Scholar]
- Wang, Z.; Song, M.; Zhang, Z.; Song, Y.; Wang, Q.; Qi, H. Beyond inferring class representatives: User-level privacy leakage from federated learning. In Proceedings of the IEEE Conference on Computer Communications (INFOCOM 2019), Paris, France, 29 April–2 May 2019; pp. 2512–2520. [Google Scholar]
- Chabanne, H.; de Wargny, A.; Milgram, J.; Morel, C.; Prouff, E. Privacy-preserving classification on deep neural network. IACR Cryptol. ePrint Arch. 2017, 2017, 35. [Google Scholar]
- Ma, X.; Ma, J.; Li, H.; Jiang, Q.; Gao, S. PDLM: Privacy-preserving deep learning model on cloud with multiple keys. IEEE Trans. Serv. Comput. 2018. [Google Scholar] [CrossRef]
- Moriai, S. Privacy-Preserving Deep Learning via Additively Homomorphic Encryption. In Proceedings of the 2019 IEEE 26th Symposium on Computer Arithmetic (ARITH), Kyoto, Japan, 10–12 June 2019; p. 198. [Google Scholar]
- Mohassel, P.; Zhang, Y. Secureml: A system for scalable privacy-preserving machine learning. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 19–38. [Google Scholar]
- Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H.B.; Mironov, I.; Talwar, K.; Zhang, L. Deep learning with differential privacy. In Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security, Vienna, Austria, 24–28 October 2016; pp. 308–318. [Google Scholar]
- Bonawitz, K.; Ivanov, V.; Kreuter, B.; Marcedone, A.; McMahan, H.B.; Patel, S.; Ramage, D.; Segal, A.; Seth, K. Practical secure aggregation for privacy-preserving machine learning. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 1175–1191. [Google Scholar]
- Liu, Y.; Ma, Z.; Liu, X.; Ma, S.; Nepal, S.; Deng, R. Boosting privately: Privacy-preserving federated extreme boosting for mobile crowdsensing. arXiv 2019, arXiv:1907.10218. [Google Scholar]
- Hao, M.; Li, H.; Xu, G.; Liu, S.; Yang, H. Towards Efficient and Privacy-Preserving Federated Deep Learning. In Proceedings of the ICC 2019-2019 IEEE International Conference on Communications (ICC), Shanghai, China, 20–24 May 2019; pp. 1–6. [Google Scholar]
- Zhao, L.; Wang, Q.; Zou, Q.; Zhang, Y.; Chen, Y. Privacy-preserving collaborative deep learning with unreliable participants. IEEE Trans. Inf. Forensics Secur. 2019, 15, 1486–1500. [Google Scholar] [CrossRef] [Green Version]
Parameters | Number of Users | Participation Ratio | Mini-Batch Size | Classes per User | Gradient Size | Number of Epochs | Privacy Budget |
---|---|---|---|---|---|---|---|
Value | Various | 10% | 20 | 10 | 32-bits | Various | 0.5 |
MNIST (Accuracy = 91.3) | CIFAR-10 (Accuracy = 87.6) | |
---|---|---|
Baseline | 2218/2218 MB | 35653 MB/35653 MB |
FedAvg | 119.65 MB/119.65 MB | 2589.5 MB/2589.5 MB |
FedAvg | 84.73 MB/84.73 MB | 1665.7 MB/1665.7 MB |
PPDL | 98.63 MB/311.6 MB | 1472.2 MB/4739.2 MB |
PPDL | 63.74 MB/432.2 MB | 958.3 MB/6342.4 MB |
FedOpt | 10.2 MB/102 MB | 109.23 MB/1090.3 MB |
FedOpt | 14.6 MB/146 MB | 172.3 MB/1723 MB |
Functionality | PSA | XGB | EPFDL | PPCL | FedOpt |
---|---|---|---|---|---|
Communication Efficient | ✓ | ✓ | |||
Collaborative Training | ✓ | ✓ | ✓ | ✓ | |
Non-IID Support | ✓ | ||||
Gradient Confidentiality | ✓ | ✓ | ✓ | ✓ | |
Attack Resilience | ✓ | ✓ | |||
Post-Quantum Security | ✓ | ✓ | |||
Collusion Resistance | ✓ | ✓ | ✓ | ✓ | |
Fast Convergence Speed | ✓ | ✓ | ✓ | ||
Application Aware | ✓ | ✓ | ✓ | ||
Algorithm Complexity | ✓ | ✓ | ✓ |
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
Share and Cite
Asad, M.; Moustafa, A.; Ito, T. FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning. Appl. Sci. 2020, 10, 2864. https://doi.org/10.3390/app10082864
Asad M, Moustafa A, Ito T. FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning. Applied Sciences. 2020; 10(8):2864. https://doi.org/10.3390/app10082864
Chicago/Turabian StyleAsad, Muhammad, Ahmed Moustafa, and Takayuki Ito. 2020. "FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning" Applied Sciences 10, no. 8: 2864. https://doi.org/10.3390/app10082864
APA StyleAsad, M., Moustafa, A., & Ito, T. (2020). FedOpt: Towards Communication Efficiency and Privacy Preservation in Federated Learning. Applied Sciences, 10(8), 2864. https://doi.org/10.3390/app10082864