Computer Science – Information Theory
Scientific paper
2008-06-16
Computer Science
Information Theory
5 pages in IEEE 2 column format, submitted to IEEE Trans. Information Theory in June, 2008
Scientific paper
In this paper, a channel that is contaminated by two independent Gaussian noises $S ~ N(0,Q)$ and $Z_0 ~ N(0,N_0)$ is considered. The capacity of this channel is computed when independent noisy versions of $S$ are known to the transmitter and/or receiver. It is shown that the channel capacity is greater then the capacity when $S$ is completely unknown, but is less then the capacity when $S$ is perfectly known at the transmitter or receiver. For example, if there is one noisy version of $S$ known at the transmitter only, the capacity is $0.5\log(1+\frac{P}{Q(N_1/(Q+N_1))+N_0})$, where $P$ is the input power constraint and $N_1$ is the power of the noise corrupting $S$. Further, it is shown that the capacity with knowledge of any independent noisy versions of $S$ at the transmitter is equal to the capacity with knowledge of the statistically equivalent noisy versions of $S$ at the receiver.
Peng Yong
Rajan Dinesh
No associations
LandOfFree
On the Capacity Equivalence with Side Information at Transmitter and Receiver does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with On the Capacity Equivalence with Side Information at Transmitter and Receiver, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and On the Capacity Equivalence with Side Information at Transmitter and Receiver will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-92491