Proximity penalty priors for Bayesian mixture models

Statistics – Methodology

Scientific paper

Rate now

  [ 0.00 ] – not rated yet Voters 0   Comments 0

Details

14 pages, 6 figures

Scientific paper

When using mixture models it may be the case that the modeller has a-priori beliefs or desires about what the components of the mixture should represent. For example, if a mixture of normal densities is to be fitted to some data, it may be desirable for components to focus on capturing differences in location rather than scale. We introduce a framework called proximity penalty priors (PPPs) that allows this preference to be made explicit in the prior information. The approach is scale-free and imposes minimal restrictions on the posterior; in particular no arbitrary thresholds need to be set. We show the theoretical validity of the approach, and demonstrate the effects of using PPPs on posterior distributions with simulated and real data.

No associations

LandOfFree

Say what you really think

Search LandOfFree.com for scientists and scientific papers. Rate them and share your experience with other people.

Rating

Proximity penalty priors for Bayesian mixture models does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.

If you have personal experience with Proximity penalty priors for Bayesian mixture models, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Proximity penalty priors for Bayesian mixture models will most certainly appreciate the feedback.

Rate now

     

Profile ID: LFWR-SCP-O-113546

  Search
All data on this website is collected from public sources. Our data reflects the most accurate information available at the time of publication.