Computer Science – Computation and Language
Scientific paper
1994-09-13
Computer Science
Computation and Language
To appear in Grammatical Inference and Applications, Second International Colloquium on Grammatical Inference; Springer Verlag
Scientific paper
We describe a framework for inducing probabilistic grammars from corpora of positive samples. First, samples are {\em incorporated} by adding ad-hoc rules to a working grammar; subsequently, elements of the model (such as states or nonterminals) are {\em merged} to achieve generalization and a more compact representation. The choice of what to merge and when to stop is governed by the Bayesian posterior probability of the grammar given the data, which formalizes a trade-off between a close fit to the data and a default preference for simpler models (`Occam's Razor'). The general scheme is illustrated using three types of probabilistic grammars: Hidden Markov models, class-based $n$-grams, and stochastic context-free grammars.
Omohundro Stephen M.
Stolcke Andreas
No associations
LandOfFree
Inducing Probabilistic Grammars by Bayesian Model Merging does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Inducing Probabilistic Grammars by Bayesian Model Merging, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Inducing Probabilistic Grammars by Bayesian Model Merging will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-190259