Title :
Dominance of Bayesian Networks and Efficient Learning of Generalized Latent Class Models
Author :
Chen, Yixin ; Hua, Dong ; Liu, Fang
Author_Institution :
Dept. of Comput. Sci., Washington Univ., St. Louis, MO
Abstract :
A major challenge for learning Bayesian networks is the complexity in searching the huge space of models and parameters. The computational cost is higher when the model topology is more flexible. In this paper, we propose the notion of dominance which can lead to strong pruning of the search space and significant reduction of learning complexity, and apply this notion to the generalized latent class (GLC) models, a class of Bayesian networks for clustering categorical data. GLC models can address the local dependence problem in latent class analysis by assuming a very general graph structure. However, The flexible topology of GLC leads to large increase of the learning complexity. We first propose the concept of dominance and related theoretical results which is general for all Bayesian networks. Based on dominance, we propose an efficient learning algorithm for GLC. A core technique to prune dominated models is regularization, which can eliminate dominated models, leading to significant pruning of the search space. Significant improvements on the modeling quality and time complexity on real datasets are reported.
Keywords :
belief networks; learning (artificial intelligence); pattern clustering; search problems; Bayesian networks; categorical data clustering; generalized latent class models; latent class analysis; learning complexity; Artificial intelligence; Bayesian methods; Computational efficiency; Computer science; Diseases; Graphical models; Learning; Network topology; Probability distribution; Solids; Bayesian networks; dominance; generalized latent class;
Conference_Titel :
Tools with Artificial Intelligence, 2008. ICTAI '08. 20th IEEE International Conference on
Conference_Location :
Dayton, OH
Print_ISBN :
978-0-7695-3440-4
DOI :
10.1109/ICTAI.2008.28