Statistics – Machine Learning
Scientific paper
2009-12-31
Statistics
Machine Learning
20 pages, 6 figures, AISTATS 2010, Revised
Scientific paper
Deep belief networks are a powerful way to model complex probability distributions. However, learning the structure of a belief network, particularly one with hidden units, is difficult. The Indian buffet process has been used as a nonparametric Bayesian prior on the directed structure of a belief network with a single infinitely wide hidden layer. In this paper, we introduce the cascading Indian buffet process (CIBP), which provides a nonparametric prior on the structure of a layered, directed belief network that is unbounded in both depth and width, yet allows tractable inference. We use the CIBP prior with the nonlinear Gaussian belief network so each unit can additionally vary its behavior between discrete and continuous representations. We provide Markov chain Monte Carlo algorithms for inference in these belief networks and explore the structures learned on several image data sets.
Adams Ryan Prescott
Ghahramani Zoubin
Wallach Hanna M.
No associations
LandOfFree
Learning the Structure of Deep Sparse Graphical Models does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Learning the Structure of Deep Sparse Graphical Models, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Learning the Structure of Deep Sparse Graphical Models will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-104795