Title :
Taking on the curse of dimensionality in joint distributions using neural networks
Author :
Bengio, Samy ; Bengio, Yoshua
Author_Institution :
CIRAN, CIRANO, Montreal, Que., Canada
fDate :
5/1/2000 12:00:00 AM
Abstract :
The curse of dimensionality is severe when modeling high-dimensional discrete data: the number of possible combinations of the variables explodes exponentially. We propose an architecture for modeling high-dimensional data that requires resources (parameters and computations) that grow at most as the square of the number of variables, using a multilayer neural network to represent the joint distribution of the variables as the product of conditional distributions. The neural network can be interpreted as a graphical model without hidden random variables, but in which the conditional distributions are tied through the hidden units. The connectivity of the neural network can be pruned by using dependency tests between the variables (thus reducing significantly the number of parameters). Experiments on modeling the distribution of several discrete data sets show statistically significant improvements over other methods such as naive Bayes and comparable Bayesian networks and show that significant improvements can be obtained by pruning the network
Keywords :
graph theory; multilayer perceptrons; probability; conditional distributions; dependency tests; dimensionality; graphical model; high-dimensional discrete data; joint distributions; multilayer neural network; network pruning; Computer architecture; Context modeling; Data mining; Graphical models; Intelligent networks; Multi-layer neural network; Neural networks; Random variables; Smoothing methods; Testing;
Journal_Title :
Neural Networks, IEEE Transactions on