Statistics – Machine Learning
Scientific paper
2011-11-25
Statistics
Machine Learning
Submitted
Scientific paper
This paper addresses the estimation of the latent dimensionality in nonnegative matrix factorization (NMF) with the \beta-divergence. The \beta-divergence is a family of cost functions that includes the squared Euclidean distance, Kullback-Leibler and Itakura-Saito divergences as special cases. Learning the model order is important as it is necessary to strike the right balance between data fidelity and overfitting. We propose a Bayesian model based on automatic relevance determination in which the columns of the dictionary matrix and the rows of the activation matrix are tied together through a common scale parameter in their prior. A family of majorization-minimization algorithms is proposed for maximum a posteriori (MAP) estimation. A subset of scale parameters is driven to a small lower bound in the course of inference, with the effect of pruning the corresponding spurious components. We demonstrate the efficacy and robustness of our algorithms by performing extensive experiments on synthetic data, the swimmer dataset, a music decomposition example and a stock price prediction task.
Févotte Cédric
Tan Vincent Y. F.
No associations
LandOfFree
Automatic Relevance Determination in Nonnegative Matrix Factorization with the β-Divergence does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Automatic Relevance Determination in Nonnegative Matrix Factorization with the β-Divergence, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Automatic Relevance Determination in Nonnegative Matrix Factorization with the β-Divergence will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-175006