×

Stochastic privacy-preserving methods for nonconvex sparse learning. (English) Zbl 07841851

Summary: Sparse learning is essential in mining high-dimensional data. Iterative hard thresholding (IHT) methods are effective for optimizing nonconvex objectives for sparse learning. However, IHT methods are vulnerable to adversary attacks that infer sensitive data. Although pioneering works attempted to relieve such vulnerability, they confront the issue of high computational cost for large-scale problems. We propose two differentially private stochastic IHT: one based on the stochastic gradient descent method (DP-SGD-HT) and the other based on the stochastically controlled stochastic gradient method (DP-SCSG-HT). The DP-SGD-HT method perturbs stochastic gradients with small Gaussian noise rather than full gradients, which are computationally expensive. As a result, computational complexity is reduced from \(O(n\log(n))\) to a lower \(O(b\log(n))\), where \(n\) is the sample size and \(b\) is the mini-batch size used to compute stochastic gradients. The DP-SCSG-HT method further perturbs the stochastic gradients controlled by large-batch snapshot gradients to reduce stochastic gradient variance. We prove that both algorithms guarantee differential privacy and have linear convergence rates with estimation bias. A utility analysis examines the relationship between convergence rate and the level of perturbation, yielding the best-known utility bound for nonconvex sparse optimization. Extensive experiments show that our algorithms outperform existing methods.

MSC:

68-XX Computer science
65-XX Numerical analysis

Software:

TensorFlow; RAPPOR; MNIST
Full Text: DOI

References:

[1] M. Abadi, A. Agarwal, P. Barham, E. Brevdo, Z. Chen, C. Citro, G.S. Corrado, A. Davis, J. Dean, M. Devin, et al. Tensorflow: Large-scale machine learning on heterogeneous distributed systems. arXiv preprint arXiv:1603.04467, 2016a.
[2] Abadi, M.; Chu, A.; Goodfellow, I.; McMahan, H. B.; Mironov, I.; Talwar, K.; Zhang, L., Deep learning with differential privacy, (Proceedings of the 2016 ACM SIGSAC Conference on Computer and Communications Security (2016), ACM), 308-318
[3] Adnan, M.; Kalra, S.; Cresswell, J. C.; Taylor, G. W.; Tizhoosh, H. R., Federated learning and differential privacy for medical image analysis, Scientific reports, 12, 1, 1-10 (2022)
[4] Bahmani, S.; Raj, B.; Boufounos, P. T., Greedy sparsity-constrained optimization, Journal of Machine Learning Research, 14, Mar, 807-841 (2013) · Zbl 1320.90046
[5] Bassily, R.; Smith, A.; Thakurta, A., Private empirical risk minimization: Efficient algorithms and tight error bounds, (2014 IEEE 55th Annual Symposium on Foundations of Computer Science (2014), IEEE), 464-473
[6] Blumensath, T.; Davies, M. E., Iterative hard thresholding for compressed sensing, Applied and computational harmonic analysis, 27, 3, 265-274 (2009) · Zbl 1174.94008
[7] Chaudhuri, K.; Monteleoni, C.; Sarwate, A. D., Differentially private empirical risk minimization, Journal of Machine Learning Research, 12, Mar, 1069-1109 (2011) · Zbl 1280.62073
[8] A.E.C. Cloud. Amazon web services. Retrieved November, 9 (2011): 2011, 2011.
[9] Deng, L., The mnist database of handwritten digit images for machine learning research [best of the web], IEEE Signal Processing Magazine, 29, 6, 141-142 (2012)
[10] Dwork, C.; Kenthapadi, K.; McSherry, F.; Mironov, I.; Naor, M., Our data, ourselves: Privacy via distributed noise generation, (Annual International Conference on the Theory and Applications of Cryptographic Techniques (2006), Springer), 486-503 · Zbl 1140.94336
[11] Dwork, C.; McSherry, F.; Nissim, K.; Smith, A., Calibrating noise to sensitivity in private data analysis, (Theory of cryptography conference (2006), Springer), 265-284 · Zbl 1112.94027
[12] Dwork, C.; Rothblum, G. N.; Vadhan, S., Boosting and differential privacy, (2010 IEEE 51st Annual Symposium on Foundations of Computer Science (2010), IEEE), 51-60
[13] El Ouadrhiri, A.; Abdelhadi, A., Differential privacy for deep and federated learning: A survey, IEEE Access, 10, 22359-22380 (2022)
[14] M. Elibol, L. Lei, and M.I. Jordan. Variance reduction with sparse gradients. arXiv preprint arXiv:2001.09623, 2020.
[15] Erlingsson, Ú.; Pihur, V.; Korolova, A., Rappor: Randomized aggregatable privacy-preserving ordinal response, (Proceedings of the 2014 ACM SIGSAC conference on computer and communications security (2014)), 1054-1067
[16] Guo, J.; Ding, X.; Wang, T.; Jia, W., Combinatorial resources auction in decentralized edge-thing systems using blockchain and differential privacy, Information Sciences (2022) · Zbl 1539.68043
[17] R.B. Harikandeh, M.O. Ahmed, A. Virani, M. Schmidt, J. Konečný, and S. Sallinen. Stopwasting my gradients: Practical svrg. In Advances in Neural Information Processing Systems, pages 2251-2259, 2015.
[18] P. Jain, A. Tewari, and P. Kar. On iterative hard thresholding methods for high-dimensional m-estimation. In Advances in Neural Information Processing Systems, pages 685-693, 2014.
[19] Jiang, X.; Niu, C.; Ying, C.; Wu, F.; Luo, Y., Pricing gan-based data generators under rényi differential privacy, Information Sciences, 602, 57-74 (2022) · Zbl 1533.91300
[20] R. Johnson and T. Zhang. Accelerating stochastic gradient descent using predictive variance reduction. In Advances in neural information processing systems, pages 315-323, 2013.
[21] Kasiviswanathan, S. P.; Lee, H. K.; Nissim, K.; Raskhodnikova, S.; Smith, A., What can we learn privately?, SIAM Journal on Computing, 40, 3, 793-826 (2011) · Zbl 1235.68093
[22] Kermany, D. S.; Goldbaum, M.; Cai, W.; Valentim, C. C.; Liang, H.; Baxter, S. L.; McKeown, A.; Yang, G.; Wu, X.; Yan, F., Identifying medical diagnoses and treatable diseases by image-based deep learning, Cell, 172, 5, 1122-1131 (2018)
[23] D. Kifer, A. Smith, and A. Thakurta. Private convex empirical risk minimization and high-dimensional regression. In Conference on Learning Theory, pages 25-1, 2012.
[24] L. Lei and M. Jordan. Less than a single pass: Stochastically controlled stochastic gradient. In Artificial Intelligence and Statistics, pages 148-156, 2017.
[25] T. Li, A.K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar, and V. Smith. Federated optimization in heterogeneous networks. arXiv preprint arXiv:1812.06127, 2018.
[26] X. Li, R. Arora, H. Liu, J. Haupt, and T. Zhao. Nonconvex sparse learning via stochastic optimization with progressive variance reduction. arXiv preprint arXiv:1605.02711, 2016a.
[27] Li, X.; Li, H.; Zhu, H.; Huang, M., The optimal upper bound of the number of queries for laplace mechanism under differential privacy, Information Sciences, 503, 219-237 (2019) · Zbl 1453.68072
[28] X. Li, T. Zhao, R. Arora, H. Liu, and J. Haupt. Stochastic variance reduced optimization for nonconvex sparse learning. In International Conference on Machine Learning, pages 917-925, 2016b.
[29] G. Liang, Q. Tong, C.J. Zhu, and J. Bi. An effective hard thresholding method based on stochastic variance reduction for nonconvex sparse learning. In AAAI, pages 1585-1592, 2020.
[30] Mallat, S. G.; Zhang, Z., Matching pursuits with time-frequency dictionaries, IEEE Transactions on signal processing, 41, 12, 3397-3415 (1993) · Zbl 0842.94004
[31] Mironov, I., Rényi differential privacy, (2017 IEEE 30th Computer Security Foundations Symposium (CSF) (2017), IEEE), 263-275
[32] J. Near. Differential privacy at scale: Uber and berkeley collaboration. In Enigma 2018 (Enigma 2018), 2018.
[33] Nguyen, N.; Needell, D.; Woolf, T., Linear convergence of stochastic iterative greedy algorithms with sparse constraints, IEEE Transactions on Information Theory, 63, 11, 6869-6895 (2017) · Zbl 1390.94326
[34] Pati, Y. C.; Rezaiifar, R.; Krishnaprasad, P. S., Orthogonal matching pursuit: Recursive function approximation with applications to wavelet decomposition, (Proceedings of 27th Asilomar conference on signals, systems and computers (1993), IEEE), 40-44
[35] A. Rényi et al. On measures of entropy and information. In Proceedings of the Fourth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory of Statistics. The Regents of the University of California, 1961. · Zbl 0106.33001
[36] K. Talwar, A.G. Thakurta, and L. Zhang. Nearly optimal private lasso. In Advances in Neural Information Processing Systems, pages 3025-3033, 2015.
[37] Q. Tong, G. Liang, T. Zhu, and J. Bi. Federated nonconvex sparse learning. arXiv preprint arXiv:2101.00052, 2020.
[38] Truex, S.; Liu, L.; Gursoy, M. E.; Yu, L.; Wei, W., Demystifying membership inference attacks in machine learning as a service, IEEE Transactions on Services Computing (2019)
[39] Wang, D.; Xu, J., On sparse linear regression in the local differential privacy model, (International Conference on Machine Learning (2019)), 6628-6637
[40] D. Wang, M. Ye, and J. Xu. Differentially private empirical risk minimization revisited: Faster and more general. In Advances in Neural Information Processing Systems, pages 2722-2731, 2017.
[41] Wang, H.; Wang, H., Correlated tuple data release via differential privacy, Information Sciences, 560, 347-369 (2021) · Zbl 1484.68061
[42] L. Wang and Q. Gu. Differentially private iterative gradient hard thresholding for sparse learning. In 28th International Joint Conference on Artificial Intelligence, 2019a.
[43] L. Wang and Q. Gu. A knowledge transfer framework for differentially private sparse learning. arXiv preprint arXiv:1909.06322, 2019b.
[44] L. Wang, B. Jayaraman, D. Evans, and Q. Gu. Efficient privacy-preserving nonconvex optimization. arXiv preprint arXiv:1910.13659, 2019.
[45] Y.-X. Wang, B. Balle, and S. Kasiviswanathan. Subsampled r⧹)ényi differential privacy and analytical moments accountant. arXiv preprint arXiv:1808.00087, 2018.
[46] Wu, X.; Li, F.; Kumar, A.; Chaudhuri, K.; Jha, S.; Naughton, J., Bolt-on differential privacy for scalable stochastic gradient descent-based analytics, (Proceedings of the 2017 ACM International Conference on Management of Data (2017)), 1307-1322
[47] P. Zhou, X. Yuan, and J. Feng. Efficient stochastic gradient hard thresholding. In Advances in Neural Information Processing Systems, pages 1988-1997, 2018.
[48] Zhu, J.; Fang, X.; Guo, Z.; Niu, M. H.; Cao, F.; Yue, S.; Liu, Q. Y., Ibm cloud computing powering a smarter planet, (IEEE International Conference on Cloud Computing (2009), Springer), 621-625
This reference list is based on information provided by the publisher or from digital mathematics libraries. Its items are heuristically matched to zbMATH identifiers and may contain data conversion errors. In some cases that data have been complemented/enhanced by data from zbMATH Open. This attempts to reflect the references listed in the original paper as accurately as possible without claiming completeness or a perfect matching.