DocumentCode
3724136
Title
DRN: Bringing Greedy Layer-Wise Training into Time Dimension
Author
Xiaoyi Li;Xiaowei Jia;Hui Li;Houping Xiao;Jing Gao;Aidong Zhang
Author_Institution
Dept. of Comput. Sci. &
fYear
2015
Firstpage
859
Lastpage
864
Abstract
Sequential data modeling has received growing interests due to its impact on real world problems. Sequential data is ubiquitous - financial transactions, advertise conversions and disease evolution are examples of sequential data. A long-standing challenge in sequential data modeling is how to capture the strong hidden correlations among complex features in high volumes. The sparsity and skewness in the features extracted from sequential data also add to the complexity of the problem. In this paper, we address these challenges from both discriminative and generative perspectives, and propose novel stochastic learning algorithms to model nonlinear variances from static time frames and their transitions. The proposed model, Deep Recurrent Network (DRN), can be trained in an unsupervised fashion to capture transitions, or in a discriminative fashion to conduct sequential labeling. We analyze the conditional independence of each functional module and tackle the diminishing gradient problem by developing a two-pass training algorithm. Extensive experiments on both simulated and real-world dynamic networks show that the trained DRN outperforms all baselines in the sequential classification task and obtains excellent performance in the regression task.
Keywords
"Training","Hidden Markov models","Data models","Computational modeling","Mathematical model","Data mining","Heuristic algorithms"
Publisher
ieee
Conference_Titel
Data Mining (ICDM), 2015 IEEE International Conference on
ISSN
1550-4786
Type
conf
DOI
10.1109/ICDM.2015.60
Filename
7373402
Link To Document