Physics – Condensed Matter – Statistical Mechanics
Scientific paper
2006-11-11
Physics
Condensed Matter
Statistical Mechanics
16 pages, 1 diagram (some references added)
Scientific paper
We show how Fisher's information already known particular character as the fundamental information geometric object which plays the role of a metric tensor for a statistical differential manifold, can be derived in a relatively easy manner through the direct application of a generalized logarithm and exponential formalism to generalized information-entropy measures. We shall first shortly describe how the generalization of information-entropy measures naturally comes into being if this formalism is employed and recall how the relation between all the information measures is best understood when described in terms of a particular logarithmic Kolmogorov-Nagumo average. Subsequently, extending Kullback-Leibler's relative entropy to all these measures defined on a manifold of parametrized probability density functions, we obtain the metric which turns out to be the Fisher information matrix elements times a real multiplicative deformation parameter. The metrics independence from the non-extensive character of the system, and its proportionality to the rate of change of the multiplicity under a variation of the statistical probability parameter space, emerges naturally in the frame of this representation.
No associations
LandOfFree
Generalized information-entropy measures and Fisher information does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Generalized information-entropy measures and Fisher information, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Generalized information-entropy measures and Fisher information will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-406061