Mathematics – Probability
Scientific paper
May 2001
adsabs.harvard.edu/cgi-bin/nph-data_query?bibcode=2001aipc..568....3g&link_type=abstract
BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING: 20th International Workshop. AIP Conference Proceedi
Mathematics
Probability
Information Theory And Communication Theory, Probability Theory
Scientific paper
A new derivation is presented of maximum entropy, which is an extremizing principle for assigning probability distributions from expectation values. The additive form ΣiΦ(pi) for the maximand is first proved by requiring that, when some probabilities are given, the procedure for finding the remaining probabilities should not depend on the values of the given probabilities. This condition induces functional equations whose solution generates the additive form. To find the function φ we assign two distributions in separate spaces from separate expectation values; then assign a joint distribution by taking these same values to be expectations of its marginals; then require these marginals to be the same as the separately assigned distributions. The resulting functional equations have only one viable solution-the entropic form φ(z)=-z ln z. The exploitation of marginal distributions is due to Shore and Johnson [1], but the present derivation uses weaker axioms that require only consistency with the sum and product rules. In contrast to the information-theoretic derivation of Shannon [2], no interpretation of the maximand functional is involved. .
No associations
LandOfFree
Maximum entropy from the laws of probability does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Maximum entropy from the laws of probability, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Maximum entropy from the laws of probability will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-924000