Learning word dependencies in text by means of a deep recurrent belief network

作者:

Highlights:

摘要

We propose a deep recurrent belief network with distributed time delays for learning multivariate Gaussians. Learning long time delays in deep belief networks is difficult due to the problem of vanishing or exploding gradients with increase in delay. To mitigate this problem and improve the transparency of learning time-delays, we introduce the use of Gaussian networks with time-delays to initialize the weights of each hidden neuron. From our knowledge of time delays, it is possible to learn the long delays from short delays in a hierarchical manner. In contrast to previous works, here dynamic Gaussian Bayesian networks over training samples are evolved using Markov Chain Monte Carlo to determine the initial weights of each hidden layer of neurons. In this way, the time-delayed network motifs of increasing Markov order across layers can be modeled hierarchically using a deep model. To validate the proposed Variable-order Belief Network (VBN) framework, it is applied for modeling word dependencies in text. To explore the generality of VBN, it is further considered for a real-world scenario where the dynamic movements of basketball players are modeled. Experimental results obtained showed that the proposed VBN could achieve over 30% improvement in accuracy on real-world scenarios compared to the state-of-the-art baselines.

论文关键词:Deep belief networks,Time-delays,Variable-order,Gaussian networks,Markov Chain Monte Carlo, BN=,Bayesian Network, CD=,Contrastive Divergence, DBN=,Deep Belief Network, GN=,Gaussian Network, RNN=,Recurrent Neural Network, VBN=,Variable-order Belief Network, MCMC=,Markov Chain Monte Carlo, ML=,Maximum Likelihood, MVAR=,Multivariate Autoregression, RBM=,Restricted Boltzmann Machine

论文评审过程:Received 16 November 2015, Revised 8 July 2016, Accepted 12 July 2016, Available online 13 July 2016, Version of Record 12 August 2016.

论文官网地址:https://doi.org/10.1016/j.knosys.2016.07.019