Characterizing Source-Channel Diversity Approaches Beyond the

Characterizing Source-Channel Diversity Approaches
Beyond the Distortion Exponent ∗
Brian Dunn and J. Nicholas Laneman
University of Notre Dame
Dept. of Electrical Engineering
Notre Dame, IN 46556
{bdunn2, jnl}@nd.edu
Abstract
In this paper we analytically characterize the performance of several sourcechannel diversity approaches used to communicate over single and parallel quasistatic fading channels. An approximation is developed that provides a more accurate characterization than considering only the so-called “distortion exponent”.
Specifically, we consider transmitting i.i.d. samples of a Gaussian source over quasistatic Rayleigh fading channels with receive, but not transmit, channel state information (CSI). The objective is to to minimize the mean-square distortion averaged
over channel realizations.
For a single channel, we obtain approximations for uncoded analog transmission,
single and dual-layer digital schemes, and derive a lower bound on the expected
distortion based upon CSI at the transmitter. For parallel channels, in addition to
extending the above approaches, we develop a hybrid digital-analog scheme as a
simple form of multiple descriptions and show that it offers good performance for
a wide range of channel bandwidths.
1
Introduction
In recent years it has become of increasing interest to communicate multimedia information over wireless links. Example applications include sensor data over a wireless sensor
network, voice over a cellular network, digital radio broadcasts, or audio to wireless speakers. Over the past fifty years, fundamental techniques in digital communications have
relied upon concepts outlined by Shannon in his landmark paper [8]. Unfortunately, for
theoretically optimal performance these techniques require large delays and impractical
complexity, and therefore do not explicitly apply to non-ergodic wireless link models that
result from limited channel variations and practical delay constraints. For these scenarios, the source-channel diversity framework of [5] allows for evaluation and comparison
of various communication architectures. For a thorough review of related literature, the
reader is referred to [5].
At a high level, source-channel diversity models scenarios in which ergodic sources are
communicated over non-ergodic channels. Instead of complete joint design of source and
This work has been supported in part by the National Science Foundation through contract ECS0329766.
∗
channel coding, source coding blocks and channel coding blocks are combined in various
ways, with rate, power, and bandwidth optimizations appropriate to each particular
combination. Asymptotic performance in terms of rate-distortion regions are used to
characterize source coding blocks, and capacity-versus-outage, or outage probability [6],
is used to characterize performance of channel coding blocks.
For a variety of approaches based upon single and multiple description source coding
over parallel channels, [5] shows that end-to-end average distortion behaves for large SNR
·
as1 E[D] = SNR−∆(L) , where SNR is the average channel signal-to-noise ratio, ∆(L) is the
so-called “distortion exponent”, and L is the number of channel uses per source sample.
In this paper, we refine the performance analysis to be of the form
E[D] ≈ C(L) · log(L · SNR)p · SNR−∆(L) ,
(1)
where p ∈ {0, 1, 2} and C(L) is a coefficient independent of SNR, and analyze new
approaches. Specifically, Section 2 summarizes approximations of the form (1) for a single
quasi-static fading channel with the following approaches: uncoded analog transmission,
a digital system designed using separation, and a superposition successive refinement
scheme. These approaches are compared to a lower bound based upon transmit channel
state information (CSI). We note that the distortion exponents presented in Section 2
have been found concurrently in [4]. Section 3 extends these results to parallel quasistatic fading channels and introduces a hybrid digital analog (HDA) scheme that performs
well compared to the approaches considered.
2
Communication over a Single Channel
We consider transmitting i.i.d. samples of a unit-variance Gaussian source over a single
quasi-static Rayleigh fading channel. We assume a delay constraint prevents us from
increasing the block length sufficiently to average over variations in the channel, but the
block length is large enough to approach the rate-distortion function of the source. For
regimes in the latter cannot be achieved, our results serve a lower bounds on average
distortion.
For a single block, the quasi-static fading channel is modeled as
y (n) = a · x(n) + z(n),
(2)
where the fading coefficient a is a circularly symmetric complex Gaussian random variable (RV) that remains constant over the entire block and is chosen independently between blocks. The channel inputs and outputs are xi and yi , respectively, and wi is i.i.d.
zero-mean, unit-variance circularly symmetric complex additive white Gaussian noise
(AWGN). The encoder maps K real source samples to N real channel inputs, or equivalently, to N/2 complex channel inputs. We characterize the relation between the channel
and source bandwidths through the bandwidth expansion ratio L := N/K.2
1
·
Throughout the paper, the notations f (x) = xν and f (x) ≈ g(x) correspond to the limiting operations log f (x)/ log x → ν and f (x)/g(x) → 1, respectively, under the appropriate limit with respect to
x.
2
In [5] the encoder maps K real source samples into to Nc = N/2 complex channel uses. Thus, the
bandwidth expansion ratio ω = Nc /K in [5] satisfies ω = L/2.
2.1
A Lower Bound
We derive a lower bound on the achievable distortion by considering the case in which
the transmitter has complete knowledge of the realized fading coefficient. A system that
does not have channel state information at the transmitter can do no better than one
that does, allowing us to compute a simple lower bound. Conditioned on knowledge of
a, the channel becomes an AWGN channel with known SNR. Here separation is optimal
if the source and channel code can both be adapted. This allows the transmitter to find
the realized mutual information, I (x; y ) = 12 log (1 + |a|2 SNR), and hence the maximal
rate at which the channel realization can support reliable communication. For a specific
a, the expected distortion can be found by evaluating the distortion rate function of the
source at a rate equal to the realized mutual information. Averaging over the exponential
distribution of |a|2 yields
Z
e1/SNR ∞ e−t/SNR
E[D] =
dt.
(3)
SNR 1
tL
For asymptotically high SNR, (3) can be approximated as
(
log(SNR) · SNR−1 , for L = 1
E[D] ≈
.
1
SNR−1 ,
for L > 1
L−1
2.2
(4)
Uncoded Transmission
It is well known that, for L = 1, uncoded transmission over an AWGN channel is optimal
for all values of SNR. Since uncoded transmission does not require knowledge of the
channel’s effective SNR at the transmitter, it remains optimal for the quasi-static fading
channel with L = 1. For L < 1, only the first N source samples will be transmitted, and
thus for asymptotically high SNR, the expected distortion will approach σs2 · (1 −L). This
implies for L < 1 the expected distortion can be approximated as E[D] ≈ σs2 · (1 − L).
We assume for L > 1 the extra bandwidth of N − K channel uses is not used, but
the additional power available is used for the first K channel uses, to maintain the
appropriate average power. Performing minimum mean-square error (MMSE) estimation
at the receiver results in
1
1 1/SNR
e
E1
.
(5)
E[D] =
SNR
SNR
For asymptotically high SNR, (5) can be approximated as
E[D] ≈
1
log(L · SNR) · SNR−1 .
L
(6)
We see that for L ≥ 1, uncoded transmission achieves the optimal distortion exponent
found in Section 2.1 of ∆(L) = 1. In the special case of L = 1 the expected distortion of
uncoded transmission is optimal for all SNR, however, this is not the case for L 6= 1.
2.3
Rate-Optimized Digital
We analyze the performance of standard (separate source and channel coding) digital
transmission over the quasi-static Rayleigh fading channel given by (2). Without knowledge of the fading coefficients at the transmitter, there is a non-zero probability the
realized channel mutual information will not support reliable communication at a target
rate R. This probability of outage can be computed as follows:
2R/L
e
−1
Pout (R, SNR) := Pr [I (x; y ) < R/L] = 1 − exp −
,
(7)
SNR
where R is the source rate in nats per real source sample. Using the total probability
law the expected distortion is found to be
2R/L
2R/L
e
−1
e
−1
−2R
E[D] = 1 − exp −
+e
· exp −
.
(8)
SNR
SNR
Note that the probability of outage, and hence the expected distortion, is a function
of both R and SNR. For asymptotically high SNR, the optimal rate can be approximated
as r0 + r log(SNR), and the expected distortion can be approximated as
L
E[D] ≈ (L + 1)(L · SNR)− L+1 .
(9)
From a higher level, the sub-optimality of separate source and channel coding is due to
the inherent nature of the separation. The source coder is designed under the assumption
that it’s output will be available to the source decoder with no errors, a condition that
may be impossible to meet for certain channels. When errors are present in the decoded
bit stream the source decoder may fail completely, resulting in a mean-square error equal
to the source variance. One possible solution may be to design encoding schemes that
degrade more gracefully as the quality of the channel decreases.
2.4
Superposition Successive Refinement Coding
In order to partially combat the on-off nature of rate-optimized digital transmission
that results in suboptimal performance, we consider a layered scheme using successive
refinement source coding [2]. We consider a dual-layer successive refinement code, where
the refinement layer is superimposed on the base layer and power allocation between the
layers is performed to minimize the expected distortion. The transmitted signal is the
sum of the two layers,
x = α · SNR · xB + (1 − α) · SNR · xE .
(10)
The decoding is performed as follows: The receiver first attempts to decode the
base layer xB treating the enhancement layer xE as additive noise. If the base layer
is successfully decoded, the receiver subtracts its estimate of the transmitted codeword
from the received signal and attempts to decode the enhancement layer. The average
distortion as a function of α, RB , and RE can be expressed as
E[D] = Pr[Bout ] + e−2RB · Pr[Bout ] Pr[Eout |Bout ] + e−2(RB +RE ) · Pr[Bout ] Pr[Eout |Bout ], (11)
where Bout and Eout denote the events of a base layer and enhancement layer outage,
and RB and RE denote the base and enhancement layer rates in nats per source sample,
respectively. The probability of the outage events are computed as
e2RB /L − 1
Pr[Bout ] = 1 − exp −
, and
(12)
SNR [1 − (1 − α)e2RB /L ]
e2RB /L − 1
e2RE /L − 1
Pr Eout |Bout = 1 − exp
−
.
(13)
SNR [1 − (1 − α)e2RB /L ] (1 − α)SNR
By performing the optimization over RB , RE , and α numerically, we observe that
for high SNR the optimal rates can be approximated as RB ≈ r0B + rB · log(SNR) and
RE ≈ r0E + rE · log(SNR). Also, the optimal α behaves as α ≈ 1 − SNR−α̂ , where the
constant α̂ determines the exponential rate at which more power is allocated to the base
layer. Analysis for asymptotically high SNR and minimization over r0B , rB , r0E , rE , and
α̂ yields the approximation for expected distortion
− L(L+2)
2
E[D] ≈ L
(L+1)
L+2
· (L + 1) L+1 · SNR
−
L(L+1)
L2 +L+1
,
(14)
where the distortion exponent given in (14) was also found in [4], and generalized to any
number of layers.
3
Communication over Parallel Channels
An attractive means to combat multipath fading is the use of multiple antennas, frequency
bands, or time slots to provide the user access to parallel channels. In this section we
present and analyze the performance of several methods to transmit an i.i.d. Gaussian
source over two independent Rayleigh block-fading AWGN channels. As in Section 2,
we assume a delay constraint prevents us from increasing the block length sufficiently to
average over variations in the channel, but the block length is large enough to reasonably
approximate the source coding distortion as the distortion-rate function of the source.
The corresponding channel model becomes
y1 (n) =a1 · x1 (n) + z1 (n)
y2 (n) =a2 · x2 (n) + z2 (n),
(15a)
(15b)
where a1 and a2 remain constant over the entire block. Furthermore, the encoder maps
K real source samples to N pairs of real channel inputs, or N/2 pairs of complex channel
inputs. As for a single channel, the bandwidth expansion ratio is denoted as L := N/K.
In this section we derive a lower bound on the mean-square error for communication
of an i.i.d. Gaussian source over parallel channels, and the corresponding upper bound on
the distortion exponent. We also analyze the performance of simple analog and digital
schemes, followed by successive refinement source coding over parallel channels. Finally,
we introduce a hybrid digital-analog scheme and evaluate its performance.
3.1
A Lower Bound
As for a single channel, we derive a lower bound on the expected distortion by assuming
the realized fading coefficients are known a priori at the transmitter. This allows the
transmitter to compute the mutual information between the inputs and outputs of the
composite channel, and determine the maximum rate at which reliable communication
can be supported. The expected distortion is found to be
Z ∞ −t/SNR 2
e2/SNR
e
E[D] =
dt ,
(16)
2
tL
SNR
1
which can be approximated for high SNR as
(
(log SNR)2 · SNR−2 ,
E[D] ≈
1
SNR−2 ,
(L−1)2
for L = 1
.
for L > 1
(17)
Although the distortion exponent derived for a single channel with perfect CSIT is achievable for L > 1 using uncoded transmission, this is not the case for parallel channels. A
scheme that can achieve the distortion exponent given in (16) without CSIT likely requires some form of joint source-channel coding. Next, we analyze several practical
schemes that do not require CSIT and compare their performance to the lower bound
derived in this section.
3.2
Uncoded Repetition
The natural extension of uncoded transmission considered in Section 2.2 to parallel channels consists of simply transmitting the source uncoded on each component channel.
Because s and y are jointly Gaussian RVs, the minimum mean-square error (MMSE)
estimate of s coincides with the linear least squares (LLS) estimate, for which closed
form expressions for the estimate and the distortion exist. The expected distortion using
vector LLS estimation is
1
−1
−2 1/SNR
,
(18)
E[D] = SNR − SNR e
E1
SNR
1
≈
· SNR−1 .
(19)
L
Note that (19) differs from (6) only in the absence of the log(L · SNR) term. This implies
uncoded transmission over a single channel and parallel channels have the same distortion
exponent.
3.3
Multi-Rate Digital Transmission
Perhaps the simplest form of digital transmission over parallel channels consists of performing separate source and channel coding and selection combining at the destination.
We allow the rates of each description, R1 and R2 , to be chosen independently, a condition slightly more general than that considered in [5] where the rates are constrained to
be equal. Without loss of generality we assume the (possibly) higher rate code is sent
over channel 1, i.e. R1 ≥ R2 .
Using techniques analogous to the outage probability analysis of Section 2.3, we find
the expected distortion and optimize the rate. Interestingly, numerical optimization of
R1 and R2 reveals that a lower average distortion is achieved for R1 6= R2 . In the limit
of high SNR, the expected distortion can be approximated as
2L+1
−
E[D] ≈ (L + 1) L+1 · (L · SNR)
L(2L+1)
(L+1)2
.
(20)
As previously mentioned, a simpler version of digital transmission with selection combining [5] constrains the rates to be equal. In the limit of high SNR, this digital repetition
scheme can be approximated as
2
L+2
2L
2
L
E[D] ≈ 1 +
· SNR− L+2 .
L
2
(21)
In [5] several other schemes were also analyzed in this context, such as optimal channel
coding diversity, which was shown to have a distortion exponent of ∆(L) = 2L/(L + 1).
3.4
Parallel Successive Refinement
A simple extension of the successive refinement scheme considered in Section 2.4 to parallel channels consists of transmitting the base and enhancement layers each on their
own independent channel. The encoder generates a dual-layer successive refinement code
consisting of a base layer description, xB , at rate RB nats per source sample and an enhancement layer description, xE , at rate RE nats per source sample. We also considered
power allocation between each component channel and found the gains in performance
are negligible.
We compute the expected distortion for a given RB and RE , and optimize numerically.
Just as for the successive refinement scheme considered for a single channel, the optimal
rates can be approximated as RB ≈ r0B + rB · log(SNR) and RE ≈ r0E + rE · log(SNR)
for high SNR. Minimizing over r0B , r0E , rB , and rE leads to the approximation
L+2
− L(L+2)
2
E[D] ≈ (L + 1) L+1 · (L · SNR)
3.5
(L+1)
.
(22)
Hybrid Digital-Analog Transmission
We consider a scheme that offers good performance for all values of L, and has the same
distortion exponent as optimal channel coding diversity [5]. The technique consists of
transmitting both an analog and a digital version of the source, one on each component
channel, and is therefore a hybrid digital-analog (HDA) scheme, with similarities to that
presented in [7]. On the first channel we transmit the source uncoded; this is optimal
in terms of its distortion exponent. The distortion on this channel will be independent
of the source, because it is solely caused by additive noise. On the second channel we
perform rate-optimized single-layer separate source and channel coding. Note that this
scheme can be thought of as a form of multiple descriptions coding [1] in the sense that
we are providing complete, yet complementary, description of the source on each channel.
We assume the receiver has perfect knowledge of the fading coefficients and can determine the reliability of decoding the digital transmission. If reliable decoding is possible,
the receiver decodes the digital transmission to form an estimate of the source ŝ2 and
linearly combines it with the noisy uncoded and scaled version ŝ1 . If we are unable to
reliably decode the digital transmission, the final source reconstruction is the MMSE
estimate of s given the noisy uncoded description. Using the LLS combiner results in an
average distortion
2R/L
1
1
1
e
−1
exp
E1
E[D] = 1 − exp −
SNR
SNR
SNR
SNR
2R/L
2R 2R e
−1 1
e
e
+ exp −
exp
E1
.
SNR
SNR
SNR
SNR
As in other digital systems studied, the numerically optimized rate grows linearly with
log(SNR) for asymptotically high SNR, and can be approximated as R ≈ r0 + r · SNR.
The final high SNR approximation is
1−L
2L
E[D] ≈ L 1+L log(L · SNR) · SNR− L+1 .
(23)
In the limit of high SNR the digital distortion can be modeled as additive Gaussian
noise independent of the source, with variance equal to the distortion-rate function of
the source [3]. In this case, the MMSE estimate of s is a linear function of ỹ1 and ỹ2 .
2
0.9
1.8
0.8
1.6
0.7
1.4
0.6
1.2
∆(L)
∆(L)
1
0.5
1
0.4
0.8
0.3
0.6
0.2
0.4
0.1
0.2
0
−10
−8
−6
−4
−2
0
2
4
Bandwidth Expansion Ratio, L (dB)
6
(a) Single Channel
8
10
0
−10
−8
−6
−4
−2
0
2
4
Bandwidth Expansion Ratio, L (dB)
6
8
10
(b) Parallel Channels
Figure 1: Distortion exponents ∆(L) as a function of the bandwidth expansion ratio, L,
in dB for a single channel (a) and parallel channels (b). At L = −4 dB from top to bottom
the curves correspond to: (a) the CSIT upper bound, two-layer successive refinement,
single-layer separate source and channel coding, and uncoded transmission, (b) the CSIT
upper bound, hybrid digital-analog transmission, multi-rate digital transmission, analog
repetition, and parallel successive refinement.
This means for asymptotically high SNR, the LLS estimate will converge to the MMSE
estimate. Since (23) is a characterization of how the system behaves at asymptotically
high SNR, it is also valid for HDA transmission utilizing a significantly more complicated
MMSE combiner.
In addition to good performance at high SNR, the HDA scheme has the desirable
property that there is never a complete outage event. This means it may have benefits
in terms of the higher order moments of the distortion. The uncoded version of the
source ensures that it is always possible to provide an estimate of the source that is
better than the naive choice of reconstructing to the source mean. A consequence of this
attribute is that the distortion has a more desirable PDF than one where a complete
outage event occurs with non-zero probability. Although this distinction is not apparent
by comparing the first moment of the distortion, for practical systems, a higher order
moment performance criterion may be of importance.
4
4.1
Discussion
System Comparison
Figure 1 shows the distortion exponents as a function of the bandwidth expansion ratio.
The average distortion for each scheme with L = 1 and L = 3 are shown in Figure 2 for
a single channel and Figure 3 for parallel channels.
4.2
Summary
This paper considers transmitting an i.i.d. Gaussian source over an AWGN channel with
Rayleigh block-fading. We introduced a more accurate characterization of a scheme’s
0
0
−2
−5
−4
−10
−6
−15
E[D] (dB)
E[D] (dB)
−8
−10
−12
−20
−25
−14
−30
−16
−35
−18
−20
0
5
10
15
20
25
30
35
−40
40
0
5
10
15
SNR (dB)
20
25
30
35
40
SNR (dB)
(a) L = 1
(b) L = 3
Figure 2: Average distortion (solid lines) and their corresponding approximations
(dashed) on a single quasi-static Rayleigh fading AWGN channel with bandwidth expansion ratio (a) L = 1 and (b) L = 3. The lower bound (Section 2.1) on expected distortion
is shown with (+), uncoded transmission (Section 2.2) with (), optimal separate source
and channel coding (Section 2.3) with (◦), and successive refinement (Section 2.4) with
(⋄).
0
0
−5
−5
−10
−10
−15
−20
E[D] (dB)
E[D] (dB)
−15
−20
−25
−30
−25
−35
−30
−40
−35
−40
−45
0
5
10
15
20
SNR (dB)
(a) L = 1
25
30
35
40
−50
0
5
10
15
20
SNR (dB)
25
30
35
40
(b) L = 3
Figure 3: Average distortion (solid lines) and their corresponding approximations
(dashed) on parallel quasi-static Rayleigh fading AWGN channels with bandwidth expansion ratio (a) L = 1 and (b) L = 3. The lower bound (Section 2.1) on distortion is
shown with (+), analog repetition (Section 3.2) with (), multi-rate digital (Section 3.3)
with (◦), parallel successive refinement (Section 3.4) with (⋄), and hybrid digital-analog
(Section 3.5) with (△).
average distortion of the form E[D] ≈ C(L) · log(L · SNR)p · SNR−∆(L) . We showed that
the distortion exponent, although always a relevant metric, does not always provide an
accurate characterization of expected distortion. In particular, the additional parameters
C(L) and p become very important for large L. One example that illustrates this notion
is the fact that uncoded transmission over a channel with L > 1 is optimal in terms of the
distortion exponent, but there is a significant gap between E[D] for uncoded transmission
and the CSIT lower bound.
Power allocation was also considered as a means to decrease distortion for schemes
that operate with multiple layers or coding rates simultaneously. Although power allocation improved performance for most schemes, the improvement was often negligible
at high SNR and may not outweigh the added system complexity and cost. The one
exception to this was the superposition successive refinement coding scheme for a single
channel considered in Section 2.4. Because the enhancement layer was treated as additive noise, it was essential that as SNR increased, more power was allocated to the base
layer.
For parallel channels, we introduced a hybrid digital-analog scheme and found that
it has the same distortion exponent as optimal channel coding [5]. Its performance,
however, still falls well below the lower bound derived in Section 3.1.
References
[1] Abbas A. El Gamal and Thomas M. Cover. Achievable Rates for Multiple Descriptions. IEEE Trans. Inform. Theory, 28(6):851–857, November 1982.
[2] W. H. R. Equitz and Thomas M. Cover. Successive refinement of information. IEEE
Trans. Inform. Theory, 37(2):269–275, March 1991.
[3] Allen Gersho and Robert M. Gray. Vector Quantization and Signal Compression.
Kluwer Academic Publishers, Boston, MA, 1992.
[4] Deniz Gunduz and Elza Erkip. Source and Channel Coding for Quasi-Static Fading
Channels. In Proc. Asilomar Conf. Signals, Systems, and Computers, Pacific Grove,
CA, October 2005. To appear.
[5] J. Nicholas Laneman, Emin Martinian, Gregory W. Wornell, and John G. Apostolopoulos. Source-Channel Diversity Approaches for Multimedia Communication.
IEEE Trans. Inform. Theory, October 2005. To appear.
[6] Lawrence H. Ozarow, Shlomo Shamai (Shitz), and Aaron D. Wyner. Information Theoretic Considerations for Cellular Mobile Radio. IEEE Trans. Vehic. Tech., 43(5):359–
378, May 1994.
[7] Shlomo Shamai (Shitz), Sergio Verdú, and Ram Zamir.
Systematic Lossy
Source/Channel Coding. IEEE Trans. Inform. Theory, 44(2):564–579, March 1998.
[8] Claude E. Shannon. A Mathematical Theory of Communication. Bell Syst. Tech. J.,
27:379–423 and 623–656, July,October 1948.