Computer Science – Information Theory
Scientific paper
2009-03-11
Computer Science
Information Theory
33 pages, 2 figures. A shorter version of this paper is to appear in IEEE Transactions on Information Theory
Scientific paper
Within the framework of linear vector Gaussian channels with arbitrary signaling, closed-form expressions for the Jacobian of the minimum mean square error and Fisher information matrices with respect to arbitrary parameters of the system are calculated in this paper. Capitalizing on prior research where the minimum mean square error and Fisher information matrices were linked to information-theoretic quantities through differentiation, closed-form expressions for the Hessian of the mutual information and the differential entropy are derived. These expressions are then used to assess the concavity properties of mutual information and differential entropy under different channel conditions and also to derive a multivariate version of the entropy power inequality due to Costa.
Palomar Daniel P.
Payaró Miquel
No associations
LandOfFree
Hessian and concavity of mutual information, differential entropy, and entropy power in linear vector Gaussian channels does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Hessian and concavity of mutual information, differential entropy, and entropy power in linear vector Gaussian channels, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Hessian and concavity of mutual information, differential entropy, and entropy power in linear vector Gaussian channels will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-123714