Autonomous Learning System Towards Mobile Intelligence
Author:
Affiliation:

Fund Project:

Science Fund for Distinguished Young Scholars of China (61725201); R&D Projects in Key Areas of Guangdong Province of China (2020B010164002)

  • Article
  • | |
  • Metrics
  • |
  • Reference [39]
  • |
  • Related [20]
  • | | |
  • Comments
    Abstract:

    How to efficiently deploy machine learning models on mobile devices has drawn a lot of attention in both academia and industry, among which the model training is a critical part. However, with increasingly public attention on data privacy and the recently adopted laws and regulations, it becomes harder for developers to collect training data from users and thus cannot train high-quality models. Researchers have been exploring approaches of training neural networks on decentralized data. Those efforts will be summarized and their limitations be pointed out. To this end, this work presents a novel neural network training paradigm on mobile devices, which distributes all training computations associated with private data on local devices and requires no data to be uploaded in any form. Such training paradigm autonomous learning is named. To deal with two main challenges of autonomous learning, i.e., limited data volume and insufficient computing power available on mobile devices, the first autonomous learning system AutLearn is designed and implemented. It incorporates the cloud (public data, pre-training)—client (private data, transfer learning) cooperation methodology and data augmentation techniques to ensure the model convergence on mobile devices. Furthermore, by utilizing a series of optimization techniques such as model compression, neural network compiler, and runtime cache reuse, AutLearn can significantly reduce the on-client training cost. Two classical scenarios of autonomous learning are implemented based on AutLearn and carried out a set of experiments. The results showed that AutLearn can train the neural networks with comparable or even higher accuracy compared to traditional centralized/federated training mode with privacy preserved. AutLearn can also significantly reduce the computational and energy cost of neural network training on mobile devices.

    Reference
    [1] Xu M, Liu J, Liu Y, Lin FX, Liu Y, Liu X. A first look at deep learning apps on smartphones. In:Proc. of the World Wide Web Conf. 2019. 2125-2136.
    [2] General Data Protection Regulation (GDPR). 2018. https://gdpr-info.eu/
    [3] McMahan HB, Moore E, Ramage D, Hampson S. Communication-efficient learning of deep networks from decentralized data. arXiv Preprint arXiv:1602.05629, 2016.
    [4] Bost R, Popa RA, Tu S, Goldwasser S. Machine learning classification over encrypted data. In:Proc. of the NDSS. 2015. 4324-4325.
    [5] Abadi M, Chu A, Goodfellow I, McMahan HB, Mironov I, Talwar K, Zhang L. Deep learning with differential privacy. In:Proc. of the 2016 ACM SIGSAC Conf. on Computer and Communications Security. 2016. 308-318.
    [6] Konečný J, McMahan HB, Ramage D, Richtárik P. Federated optimization:Distributed machine learning for on-device intelligence. arXiv Preprint arXiv:1610.02527, 2016.
    [7] Li L, Xiong H, Wang J, Xu CZ, Guo Z. SmartPC:Hierarchical pace control in real-time federated learning system. In:Proc. of the IEEE Real-time Systems Symp. (RTSS 2019). 2019.
    [8] Bonawitz K, Ivanov V, Kreuter B, Marcedone A, McMahan HB, Patel S, Ramage D, Segal A, Seth K. Practical secure aggregation for privacy-preserving machine learning. In:Proc. of the 2017 ACM SIGSAC Conf. on Computer and Communications Security. 2017. 1175-1191.
    [9] Vaidya J, Kantarcıoğlu M, Clifton C. Privacy-preserving naive bayes classification. The VLDB Journal, 2008,17(4):879-898.
    [10] Graepel T, Lauter K, Naehrig M. ML confidential:Machine learning on encrypted data. In:Proc. of the Int'l Conf. on Information Security and Cryptology 2012. Berlin, Heidelberg:Springer-Verlag, 2012. 1-21.
    [11] Nandakumar K, Ratha N, Pankanti S, Halevi S. Towards deep neural network training on encrypted data. In:Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition Workshops. 2019.
    [12] Xiong P, Zhu T, Wang X. A Survey on differential privacy and applications. Chinese Journal of Computers, 2014,37(1):101-122(in Chinese with English abstract).
    [13] Shokri R, Shmatikov V. Privacy-preserving deep learning. In:Proc. of the 22nd ACM SIGSAC Conf. on Computer and Communications Security. 2015. 1310-1321.
    [14] Bagdasaryan E, Veit A, Hua Y, Estrin D, Shmatikov V. How to backdoor federated learning. arXiv Preprint arXiv:1807.00459, 2018.
    [15] Chellappa RK, Sin RG. Personalization versus privacy:An empirical examination of the online consumer's Dilemma. Information Technology and Management, 2005,6(2/3):181-202.
    [16] Bonawitz K, Eichner H, Grieskamp W, Huba D, Ingerman A, Ivanov V, Kiddon C, Konecny J, Mazzocchi S, McMahan HB, Van Overveldt T. Towards federated learning at scale:System design. In:Proc. of the SysML. 2019.
    [17] Zhuang F, Luo P, He Q, Shi Z. Survey on transfer learning research. Ruan Jian Xue Bao/Journal of Software, 2015,26(1):26-39(in Chinese with English abstract). http://www.jos.org.cn/1000-9825/4631.htm[doi:10.13328/j.cnki.jos.004631]
    [18] Xue J, Li J, Gong Y. Restructuring of deep neural network acoustic models with singular value decomposition. In:Proc. of the Interspeech. 2013. 2365-2369.
    [19] Wei JW, Zou K. EDA:Easy data augmentation techniques for boosting performance on text classification tasks. In:Proc. of the EMNLP/IJCNLP (1). 2019. 6381-6387.
    [20] Chen T, Moreau T, Jiang Z, Zheng L, Yan E, Shen H, Cowan M, Wang L, Hu Y, Ceze L, Guestrin C. TVM:An automated end-to-end optimizing compiler for deep learning. In:Proc. of the 13th USENIX Symp. on Operating Systems Design and Implementation (OSDI 18). 2018. 578-594.
    [21] Ragan-Kelley J, Barnes C, Adams A, Paris S, Durand F, Amarasinghe S. Halide:A language and compiler for optimizing parallelism, locality, and recomputation in image processing pipelines. ACM SIGPLAN Notices, 2013,48(6):519-530.
    [22] Zerrell T, Bruestle J. Stripe:Tensor compilation via the nested polyhedral model. arXiv Preprint arXiv:1903.06498, 2019.
    [23] Lei J, Gao X, Song J, Wang XL, Song ML. Survey of deep neural network model compression. Ruan Jian Xue Bao/Journal of Software, 2018,29(2):251-266(in Chinese with English abstract). http://wwww,.jos.org.cn/1000-9825/5428.htm[doi:10.13328/j. cnki.jos.005428]
    [24] Xu M, Qian F, Zhu M, Huang F, Pushp S, Liu X. Deepwear:Adaptive local offloading for on-wearable deep learning. IEEE Trans. on Mobile Computing, 2019,19(2):314-330.
    [25] Xu M, Zhu M, Liu Y, Lin FX, Liu X. DeepCache:Principled cache for mobile deep vision. In:Proc. of the 24th Annual Int'l Conf. (MobiCom). 2018. 129-144.
    [26] Yang TJ, Howard A, Chen B, Zhang X, Go A, Sandler M, Sze V, Adam H. Netadapt:Platform-aware neural network adaptation for mobile applications. In:Proc. of the European Conf. on Computer Vision (ECCV). 2018. 285-300.
    [27] Tan M, Chen B, Pang R, Vasudevan V, Sandler M, Howard A, Le QV. MNASnet:Platform-aware neural architecture search for mobile. In:Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition. 2019. 2820-2828.
    [28] Yang TJ, Chen YH, Sze V. Designing energy-efficient convolutional neural networks using energy-aware pruning. In:Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition. 2017. 5687-5695.
    [29] Hochreiter S, Schmidhuber J. Long short-term memory. Neural Computation, 1997,9(8):1735-1780.
    [30] Caldas S, Wu P, Li T, Konečný J, McMahan HB, Smith V, Talwalkar A. Leaf:A benchmark for federated settings. arXiv Preprint arXiv:1812.01097, 2018.
    [31] Khomenko V, Shyshkov O, Radyvonenko O, Bokhan K. Accelerating recurrent neural network training using sequence bucketing and multi-GPU data parallelization. In:Proc. of the 1st IEEE Int'l Conf. on Data Stream Mining & Processing (DSMP). 2016. 100-103.
    [32] Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H. Mobilenets:Efficient convolutional neural networks for mobile vision applications. arXiv Preprint arXiv:1704.04861, 2017.
    [33] Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In:Advances in Neural Information Processing Systems. 2012. 1097-1105.
    [34] Taylor SA, Jaques N, Nosakhare E, Sano A, Picard R. Personalized multitask learning for predicting tomorrow's mood, stress, and health. IEEE Trans. on Affective Computing, 2017.
    [35] Li M, Zhang T, Chen Y, Smola AJ. Efficient mini-batch training for stochastic optimization. In:Proc. of the 20th ACM SIGKDD Int'l Conf. on Knowledge Discovery and Data Mining. 2014. 661-670.
    附中文参考文献:
    [12] 熊平,朱天清,王晓峰.差分隐私保护及其应用.计算机学报.2014,37(1):101-122.
    [17] 庄福振,罗平,何清,史忠植.迁移学习研究进展.软件学报,2015,26(1):26-39. http://www.jos.org.cn/1000-9825/4631.html[doi:10. 13328/j.cnki.jos.004631]
    [23] 雷杰,高鑫,宋杰,王兴路,宋明黎.深度网络模型压缩综述.软件学报,2018,29(2):251-266. http://www.jos.org.cn/1000-9825/5428.htm[doi:10.13328/j.cnki.jos.005428]
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

徐梦炜,刘渊强,黄康,刘譞哲,黄罡.面向移动终端智能的自治学习系统.软件学报,2020,31(10):3004-3018

Copy
Share
Article Metrics
  • Abstract:3569
  • PDF: 8004
  • HTML: 3621
  • Cited by: 0
History
  • Received:February 07,2020
  • Revised:April 04,2020
  • Online: June 11,2020
  • Published: October 06,2020
You are the first2038183Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-4
Address:4# South Fourth Street, Zhong Guan Cun, Beijing 100190,Postal Code:100190
Phone:010-62562563 Fax:010-62562533 Email:jos@iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063