Computer Science – Information Theory
Scientific paper
2010-08-12
Computer Science
Information Theory
11 pages, 4 figures; accepted for publication in Physical Review E
Scientific paper
We derive an approximate expression for mutual information in a broad class of discrete-time stationary channels with continuous input, under the constraint of vanishing input amplitude or power. The approximation describes the input by its covariance matrix, while the channel properties are described by the Fisher information matrix. This separation of input and channel properties allows us to analyze the optimality conditions in a convenient way. We show that input correlations in memoryless channels do not affect channel capacity since their effect decreases fast with vanishing input amplitude or power. On the other hand, for channels with memory, properly matching the input covariances to the dependence structure of the noise may lead to almost noiseless information transfer, even for intermediate values of the noise correlations. Since many model systems described in mathematical neuroscience and biophysics operate in the high noise regime and weak-signal conditions, we believe, that the described results are of potential interest also to researchers in these areas.
No associations
LandOfFree
Information capacity in the weak-signal approximation does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with Information capacity in the weak-signal approximation, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and Information capacity in the weak-signal approximation will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-586677