Deep Generative Neural Networks Based on Real-valued RBM with Auxiliary Hidden Units
Author:
Affiliation:

Clc Number:

TP18

Fund Project:

National Natural Science Foundation of China (61976216, 61672522)

  • Article
  • | |
  • Metrics
  • |
  • Reference [19]
  • |
  • Related [20]
  • |
  • Cited by
  • | |
  • Comments
    Abstract:

    Restricted Boltzmann machine (RBM) is a probabilistic undirected graph, and most traditional RBM models assume that their hidden layer units are binary. The advantage of binary units is their calculation process and sampling process are relatively simple. However, binarized hidden units may bring information loss to the process of feature extraction and data reconstruction. Therefore, a key research point of RBM theory is to construct real-valued visible layer units and hidden layer units, meanwhile, maintain the effectiveness of model training. In this study, the binary units are extended to real-valued units to model data and extract features. To achieve this, specifically, an auxiliary unit is added between the visible layer and the hidden layer, and then the graph regularization term is introduced into the energy function. Based on the binary auxiliary unit and graph regularization term, the data on the manifold has a higher probability to be mapped as a parameterized truncated Gaussian distribution, simultaneously, the data far from the manifold has a higher probability to be mapped as Gaussian noises. The hidden units can be sampled as real-valued units from the parameterized Gaussian distribution and Gaussian noises. In this study, the resulting RBM based model is called restricted Boltzmann machine with auxiliary units (ARBM). Moreover, the effectiveness of the proposed model is analyzed theoretically. The effectiveness of the model in image reconstruction task and image generation task is verified by experiments.

    Reference
    [1] Kingma D, Welling M. Auto-encoding variational bayes. In:Proc. of the Int'l Conf. on Learning Representations. 2014.
    [2] Goodfellow I, Pouget-Abadie J, Mirza M, et al. Generative adversarial networks. In:Proc. of the Advances in Neural Information Processing Systems. 2014.
    [3] Kuleshov V, Ermon S. Neural variational inference and learning in undirected graphical models. In:Proc. of the Advances in Neural Information Processing Systems. 2017.
    [4] Fisher C, Smith A, Walsh J. Boltzmann encoded adversarial machines. arXiv:1804.08682, arXiv, 2018.
    [5] Ranzato M, Krizhevsky A, Hinton G. Factored 3-way restricted Boltzmann machines for modeling natural images. Journal of Machine Learning Research, 2010,9:621-628.
    [6] Courville A, Desjardins G, Bergstra J, et al. The spike-and-slab RBM and extensions to discrete and sparse data distributions. IEEE Trans. on Pattern Analysis & Machine Intelligence, 2014,36(9):1874-1887.
    [7] Sohn K, Zhou G, Lee C, et al. Learning and selecting features jointly with point-wise gated Boltzmann machines. In:Proc. of the Int'l Conf. on Machine Learning. 2013.
    [8] Srivastava N, Hinton G, Krizhevsky A, et al. Dropout:A simple way to prevent neural networks from overfitting. Journal of Machine Learning Research, 2014,15(1):1929-1958.
    [9] Blundell C, Cornebise J, Kavukcuoglu K. Weight uncertainty in neural networks. In:Proc. of the Int'l Conf. on Machine Learning. 2015.
    [10] Zhang N, Ding S, Zhang J, et al. Research on point-wise gated deep networks. Applied Soft Computing, 2017,52:1210-1221.
    [11] Huang H, Toyoizumi T. Advanced mean-field theory of the restricted Boltzmann machine. Physical Review E Statistical Nonlinear & Soft Matter Physics, 2015.
    [12] Zhang N, Ding S, Zhang J, et al. An overview on restricted Boltzmann machines. Neurocomputing, 2018,275:1186-1199.
    [13] Zhang J, Ding S, Zhang N. An overview on probability undirected graphs and their applications in image processing. Neurocomputing, 2018,321:156-168.
    [14] Isola P, Zhu J, Zhou T, et al. Image-to-image translation with conditional adversarial networks. In:Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition. 2017.
    [15] Cho K, Raiko T, Ilin A. Gaussian-bernoulli deep Boltzmann machine. In:Proc. of the IEEE Int'l Joint Conf. on Neural Networks. 2014.
    [16] Su Q, Liao X, Chen C, et al. Nonlinear statistical learning with truncated gaussian graphical models. In:Proc. of the Int'l Conf. on Machine Learning. 2016.
    [17] Radford A, Metz L, Chintala S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv:1511.06434, arXiv, 2018.
    [18] Gulrajani I, Ahmed F, Arjovsky M, et al. Improved training of wasserstein GANs. In:Advances in Neural Information Processing Systems. 2017.
    [19] Dinh L, Sohl-Dickstein J, Bengio S. Density estimation using Real NVP. In:Proc. of the Int'l Conf. of Learning Research. 2016.
    Cited by
    Comments
    Comments
    分享到微博
    Submit
Get Citation

张健,丁世飞,丁玲,张成龙.基于实值RBM的深度生成网络研究.软件学报,2021,32(12):3802-3813

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:April 14,2020
  • Revised:June 05,2020
  • Online: December 02,2021
  • Published: December 06,2021
You are the first2038154Visitors
Copyright: Institute of Software, Chinese Academy of Sciences Beijing ICP No. 05046678-4
Address:4# South Fourth Street, Zhong Guan Cun, Beijing 100190,Postal Code:100190
Phone:010-62562563 Fax:010-62562533 Email:jos@iscas.ac.cn
Technical Support:Beijing Qinyun Technology Development Co., Ltd.

Beijing Public Network Security No. 11040202500063