Enhancement of Information Transmission

LETTER
Communicated by William Bialek
Enhancement of Information Transmission Efficiency by
Synaptic Failures
Mark S. Goldman∗
[email protected]
Brain and Cognitive Sciences Department, MIT, Cambridge, MA 02139, U.S.A.
Many synapses have a high percentage of synaptic transmission failures. I
consider the hypothesis that synaptic failures can increase the efficiency
of information transmission across the synapse. I use the information
transmitted per vesicle release about the presynaptic spike train as a
measure of synaptic transmission efficiency and show that this measure
can increase with the synaptic failure probability. I analytically calculate
the Shannon mutual information transmitted across two model synapses
with probabilistic transmission: one with a constant probability of vesicle release and one with vesicle release probabilities governed by the
dynamics of synaptic depression. For inputs generated by a non-Poisson
process with positive autocorrelations, both synapses can transmit more
information per vesicle release than a synapse with perfect transmission, although the information increases are greater for the depressing
synapse than for a constant-probability synapse with the same average
transmission probability. The enhanced performance of the depressing
synapse over the constant-release-probability synapse primarily reflects a
decrease in noise entropy rather than an increase in the total transmission
entropy. This indicates a limitation of analysis methods, such as decorrelation, that consider only the total response entropy. My results suggest that
synaptic transmission failures governed by appropriately tuned synaptic
dynamics can increase the information-carrying efficiency of a synapse.
1 Introduction
Synaptic transmission failures are commonly seen in recordings of single
central synapses, with the fraction of synaptic failures at a given synapse
often exceeding the fraction of successful transmissions (Stevens & Wang,
1995; Murthy, Sejnowski, & Stevens, 1997). When a synapse fails to transmit, information about the presynaptic spike train is lost. Nevertheless, the
prevalence of synaptic failures suggests that they may have a functional role,
and a variety of such roles have been proposed. Synaptic failures could re∗ Present address: Department of Physics and Program in Neuroscience, Wellesley
College, Wellesley, MA 02481, U.S.A.
c 2004 Massachusetts Institute of Technology
Neural Computation 16, 1137–1162 (2004) 1138
M. Goldman
flect the conservation of limited synaptic resources for a synapse with a
limited pool of primed, readily releasable vesicles (Murthy et al., 1997). Alternatively, synaptic failures could provide an energy-saving mechanism
that reduces the large metabolic costs of synaptic transmission (Balasubramanian, Kimber, & Berry, 2001; Laughlin, 2001; Levy & Baxter, 2002;
Schreiber, Machens, Herz, & Laughlin, 2002). Congruent with these proposals is the suggestion that synapses could serve as dynamic filters of the
presynaptic spike train, with less important spikes being filtered out while
more important spikes are transmitted (Maass & Zador, 1999; Natschlager
& Maass, 2001; Goldman, Maldonado, & Abbott, 2002). Short-term synaptic
dynamics such as synaptic depression or facilitation have been suggested as
the mechanism underlying this role. Similarly, synaptic failures associated
with synaptic depression have been suggested as a means of gain control
by which a postsynaptic neuron can respond to important information arriving along particular synapses while filtering out less interesting inputs
arriving at other synapses (Abbott, Varela, Sen, & Nelson, 1997).
Common to the above proposals is the suggestion that synaptic failures
can increase the efficiency of information transmission by a synapse. One
possible source of inefficiency in natural spike trains is temporal autocorrelations. Spike train autocorrelations have been widely observed in the
sensory areas of animals viewing natural stimuli (Dan, Atick, & Reid, 1996)
and of freely viewing animals (Baddeley et al., 1997; Goldman et al., 2002)
and indicate that information is encoded in a redundant manner (Barlow,
1961). Reduction of this redundancy has been suggested as a computational principle underlying early sensory processing (see, e.g., Attneave,
1954; Barlow, 1961; Srinivasan, Laughlin, & Dubs, 1982; Barlow & Foldiak
1989). It has been suggested that the removal of spike train autocorrelations by the synapse could produce a more efficient representation of the
encoded information (Goldman et al., 2002). However, the mechanism by
which this decorrelation was proposed, synaptic depression, is inherently
stochastic. This means that the same spike train, if presented multiple times,
will produce different sets of transmissions across a depressing synapse. In
information-theoretic terms, for a fixed number of transmissions, decorrelation leads to maximization of the total response entropy of the transmissions
but does not generally lead to a reduction in the noise entropy. In this article,
I re-explore the hypothesis that synaptic failures can increase the efficiency
of information transmission across a synapse. As a measure of information
transmission efficiency, I use the Shannon information transmitted per vesicle release by a sequence of postsynaptic vesicle releases about a sequence
of presynaptic action potentials. As a measure of information transmission
efficiency, I use the Shannon information transmitted per vesicle release by
a sequence of postsynaptic vesicle releases about a sequence of presynaptic
action potentials (see section 2). For the cases considered here, it will be seen
that the noise entropy rather than the total response entropy dominates the
behavior of this quantity.
Both the noise and the total entropies have been calculated in previous
Enhancement of Information Transmission Efficiency
1139
studies that use direct numerical calculations (de Ruyter van Steveninck,
Lewen, Strong, Koberle, & Bialek, 1997; Strong, Koberle, de Ruyter van
Steveninck, & Bialek, 1998) of information. However, these studies are typically constrained to timescales on the order of tens of milliseconds. This limitation is due to a combinatoric explosion in the number of spike train probabilities that need to be computed numerically. Spike train autocorrelations
and short-term synaptic dynamics often extend for many hundreds of milliseconds, making numerical calculations intractable for typical spike rates.
Previous studies have approximated the information transmitted across
a depressing synapse under various simplifying assumptions (Goldman,
2000; de la Rocha, Nevado, & Parga, 2002; Fuhrmann, Segev, Markram, &
Tsodyks, 2002; Goldman et al., 2002). Here, I perform an exact analytic calculation of the information transmitted across a model depressing synapse.
To make this calculation tractable, I use simplified models of a stochastic
depressing synapse and of positively autocorrelated spike trains.
Using several examples, I analyze the effects of synaptic failures, spike
train autocorrelations, and synaptic dynamics on the transmission of information across a synapse. In section 3, the information transmitted per vesicle
release by a synapse with constant probability of transmission is compared
to that transmitted by a perfectly transmitting synapse with no synaptic failures. In section 4, the information per vesicle release for a model depressing
synapse is compared to that for the perfectly transmitting synapse. In section 5.1, the depressing and constant-probability synapses are compared
directly. This last comparison gives insight into how synaptic dynamics can
increase the efficiency of information transmission across a synapse. The
results of the analytic calculations are constrained to a particular, analytically tractable synaptic model and form of spike train autocorrelations. In
section 5.2, a qualitative framework is presented for understanding more
generally how synaptic dynamics can enhance the efficiency of information
transmission across a synapse when spike trains are autocorrelated.
2 Information Calculations with Point Process Spike Trains and
Transmissions
The calculations that follow compute the Shannon mutual information
I(S; V) conveyed by a sequence of N synaptic transmissions (vesicle releases) occurring at times v = (v1 , v2 , . . . , vN ) about a presynaptic spike
train with n spikes occurring at times s ≡ (s1 , s2 , . . . , sn ):
I(S; V) = H(V) − H(V|S)
=−
P(v) log2 (P(v))
v
+
s
P(s)
v
(2.1)
(2.2)
P(v|s) log2 (P(v|s)),
where P(·) denotes probability, the random variables S and V denote the
ensembles of spike trains and vesicle releases, respectively, and the sums are
1140
M. Goldman
over all possible realizations s and v, respectively, of the random variables
S and V. Written in this form, the information is defined as the difference
between the transmission entropy H(V) and the noise entropy H(V|S) of
the train of synaptic vesicle releases. H(V) quantifies the variability in the
sequence of synaptic vesicle releases and gives the information-carrying
capacity of the vesicle releases. Subtracting H(V|S) from the transmission
entropy removes the portion of the vesicle release train variability that arises
when the presynaptic spike train is held fixed.
The presynaptic spike and postsynaptic vesicle release sequences are
modeled by point processes, with spike and release times specified with a
constant precision δT. The probabilities P(v) and P(s) are then expressed in
terms of probability densities P (v) and P (s) as P(v) = P (v)(δT)N and P(s) =
P (s)(δT)n , respectively. A consequence of this model is that it gives rise to
a term −N log2 (δT) in H(V). Correspondingly, the information per vesicle
release I(S; V)/N contains a constant term − log2 (δT) that approaches ∞ as
δT → 0. This reflects that, in this limit, each of the N vesicle releases is specified with infinitesimal resolution and has a correspondingly large capacity
to carry information. In practice, any numerical calculation of the information requires the choice of a finite time bin that defines the meaningful
precision for a given system. Since the biologically relevant precision δT is
not necessarily known, it is useful to have a measure for comparing synaptic
transmission efficiency that is independent of δT. Because measures such as
the information transmitted per unit time or energy are dependent on δT, I
choose not to use these even though they could provide interesting results
that are calculable for specific choices of δT. I instead use the difference
between the information per vesicle release transmitted by two synapses,
which is independent of δT. In section 6, I point out some qualitative distinctions between these different measures.
3 Information Transmission for a Synapse with Constant Vesicle
Release Probability
3.1 Poisson Input with Time-varying Rate. I first calculate the information I(S; V) for a synapse with constant vesicle release probability p receiving input described by a Poisson process of arbitrary rate r(t). This example
illustrates the effect of synaptic unreliability in the case that: (1) spikes are
generated independently, and (2) transmission failures are governed by a
process that has no dynamics. After the calculation, I consider the effects of
relaxing these two assumptions.
Because the transmission probability p is time independent and the input
is Poisson, H(V) and H(V|S) can be found from the infinitesimal entropies
δH(V) and δH(V|S) corresponding to a single time bin (t, t + δT), where δT
is assumed to be small. Inserting the probabilities of vesicle release, p r(t)δT,
and failure, 1 − p r(t)δT, into equation 2.2 and ignoring terms of order (δT)2
Enhancement of Information Transmission Efficiency
1141
gives
δH(V) = −pr(t)δT log2 (pr(t)δT)−(1−pr(t)δT) log2 (1−pr(t)δT)
1
≈ pr(t)δT
− log2 (pr(t)δT)
ln 2
δH(V|S) = −r(t)δT[p log2 (p) + (1 − p) log2 (1 − p)]
(3.1)
(3.2)
(3.3)
Summing the above equations over time and dividing by the average
T
number of transmissions in a time Ttotal , N = 0 total pr(t) dt, gives the entropies and information per vesicle release (here and throughout, I assume
δT is small and denote sums over time intervals of size δT by integrals with
respect to the dummy integration variable t):
H(V)
1
=
−
N
ln 2
Ttotal
0
r(t) log2 (pr(t)δT) dt
.
Ttotal
r(t) dt
0
(3.4)
H(V|S)
1−p
= − log2 (p) +
log2 (1 − p) ,
N
p
I(S; V)
1
1−p
=
+
log2 (1 − p) −
N
ln 2
p
Ttotal
0
r(t) log2 (r(t)δT) dt
.
Ttotal
r(t) dt
0
(3.5)
(3.6)
It should be noted that, although different in detail, the above derivation is
similar in form and spirit to that for the information carried by the arrival
time of a single spike (Brenner, Strong, et al., 2000).
For a synapse with perfect transmission (p = 1), the corresponding information per vesicle release equals the entropy per spike present in the
presynaptic spike train,
Ip=1 (S; V)
H(S)
1
=
=
−
N
n
ln 2
Ttotal
0
r(t) log2 (r(t)δT) dt
,
Ttotal
r(t) dt
0
(3.7)
where n is the number of spikes in the presynaptic train.
The difference Iconst /N between the information transmitted per vesicle
release by the constant probability and perfectly transmitting synapses is
independent of the rate r(t),
Iconst /N =
I(S; V) Ip=1 (S; V)
1−p
−
=
log2 (1 − p),
N
N
p
and is shown in Figure 1.
(3.8)
1142
M. Goldman
'Iconst/N (bits)
0.0
-0.5
-1.0
-1.5
0.0
0.2
0.4
0.6
0.8
1.0
Synaptic failure probability
Figure 1: Synaptic unreliability decreases information per vesicle release for a
constant-probability synapse receiving time-varying Poisson input. The information difference Iconst /N = Iconst /N − Ip=1 /N decreases monotonically with
the synaptic failure probability 1 − p. This result is independent of the Poisson
rate r(t).
One might have naively expected the information per transmission to
be independent of p for a constant-probability synapse. This would be the
case if the information transmitted were proportional to the number of
transmissions. Instead, the information per transmission decreases monotonically with the failure probability 1 − p (see Figure 1). For example, a
constant-probability synapse with p = 0.5 transmits 1 bit per transmission
less than a perfectly transmitting synapse. Mathematically, the loss of information per vesicle release derives from the contribution of the failures to the
noise entropy, expressed by the second term of equation 3.5. Qualitatively,
the information per transmission is less for the constant-probability synapse
because, in addition to reducing the number of transmissions, there is added
uncertainty associated with not knowing how many presynaptic action potentials failed to transmit and which of the many possible arrangements
of these untransmitted action potentials was actually present in the presynaptic spike train (a similar situation is described by Shannon & Weaver,
1949).
3.2 Input Generated by a Stationary Renewal Process. For Poisson input, the probability of producing a spike in any given time bin depends on
only the rate r(t) and is independent of the spiking activity in any other
time bin. I next show that when the locations of spikes depend on the locations of previous spikes, the constant-probability synapse can transmit more
information per vesicle release than a perfectly transmitting synapse. To introduce spike-spike dependencies, I model the presynaptic spike train by a
Enhancement of Information Transmission Efficiency
1143
renewal process, which generates successive interspike intervals (ISIs) independently from one another. Furthermore, for tractability of calculation,
the analyses presented here are restricted to stationary renewal processes,
which are characterized by a time-independent probability density PISI (S )
of generating an ISI of duration S .
The stationary renewal process is a generalization of the stationary (or homogeneous) Poisson process. Unlike the homogeneous Poisson process, the
stationary renewal process generally has nonzero autocorrelations. Spike
train autocorrelations are an indication of redundancy in the neural code
(Barlow, 1961). This suggests that synaptic transmission failures might increase the efficiency of information transmission by removing redundancy
from the presynaptic spike trains.
Below, I derive general expressions for H(V)/N and H(V|S)/N for a
constant-probability synapse receiving stationary renewal spike train input. I then apply these results to a model renewal spike train constructed
to have nonnegative autocorrelations. The autocorrelation function of the
stationary renewal process spike trains is derived in appendix A.
3.2.1 General Results. The probability P(s) of generating a particular nspike train s from a stationary renewal process is given by
P(s) =
n
[PISI (S,i )δT],
(3.9)
i=1
where δT gives the precision with which all spike and transmission times
are defined, and S,i denotes the duration of the ith ISI.
For a synapse with constant vesicle release probability p receiving a spike
train generated by a stationary renewal process, the intervesicle release interval (IVI) distribution PIVI (V ) is also a stationary renewal process. This
is because the transmission probabilities across the constant-probability
synapse do not depend on the locations of spikes previous to the one transmitted. Thus, the probability of generating a particular N-vesicle-release
train v is given by
P(v) =
N
[PIVI (V,i )δT],
(3.10)
i=1
where V,i denotes the duration of the ith IVI. The IVI distribution PIVI (V )
for the constant probability synapse can be derived from the ISI distribution
PISI (S ) by using the method of Laplace transforms (Cox, 1962; Cox &
Lewis, 1966; de la Rocha et al., 2002). The resulting expression, derived in
appendix B, is
ISI (u)
pP
−1
PIVI (V ) = L
,
(3.11)
ISI (u)
1 − (1 − p)P
1144
M. Goldman
denotes the Laplace transform of the distribution P , u denotes the
where P
argument of the Laplace transform, and L−1 {·} indicates the operation of
taking the inverse Laplace transform.
Inserting equation 3.10 into the definition of the transmission entropy
(equation 2.2) and noting that the integral separates into N identical terms
gives the transmission entropy per vesicle release H(V)/N as
∞
H(V)
dV PIVI (V ) log2 (PIVI (V )δT).
(3.12)
=−
N
0
Using equations 3.11 and 3.12, H(V)/N can be calculated for any spike train
generated by a stationary renewal process.
I next find a corresponding expression for the noise entropy per release
H(V|S)/N. To calculate this quantity, note that the vesicle releases across the
constant probability synapse have no history dependence. This implies that
the conditional probability P(v|s) can be written as a product of probabilities
Q(bi |S,i ) that the ith spike triggers a vesicle release, bi = 1, or does not,
bi = 0:
P(v|s) =
n
Q(bi |S,i ),
(3.13)
i=1
where the probability Q(b|S ) is given by,
1 − p, b = 0
Q(b|S ) =
p, b = 1.
(3.14)
Substituting equation 3.13 into the definition of the noise entropy (equation 2.2) gives the noise entropy per spike H(V|S)/n,
∞
H(V|S)
=−
dS PISI (S )
Q(b|S ) log2 (Q(b|S )).
(3.15)
n
0
b=0,1
Converting this to a noise entropy per vesicle release H(V|S)/N by dividing by the probability of vesicle transmission p gives, after substitution for
Q(b|S ) from equation 3.14,
(1 − p)
H(V|S)
(3.16)
= − log2 p +
log2 (1 − p) .
N
p
This expression is independent of the ISI distribution PISI (S ) and is identical to that derived previously for the Poisson spike train, equation 3.5.
From the general expressions given above for H(V)/N (equation 3.12
using PIVI (V ) from equation 3.11) and H(V|S)/N (equation 3.16), the information per vesicle release I(S; V)/N can be obtained for any spike train
defined by a stationary renewal process.
Enhancement of Information Transmission Efficiency
1145
3.2.2 Input with Nonnegative Autocorrelations That Decay Exponentially.
In this section, the general expressions derived above are evaluated for
a stationary renewal spike train of average rate r and with positive autocorrelations of the form C(τ ) = Ae−|τ |/τcorr . This example allows independent
study of the effects of changing the average rate r, the magnitude of autocorrelations A (normalized to be independent of r), and the exponential
decay time of the autocorrelations τcorr .
In appendix A, I show that autocorrelations of this form are generated
when the ISI distribution is given by (Cox, 1962; Cox & Lewis, 1966; de la
Rocha et al., 2002)
PISI (S ) = αλ− e−λ− S + (1 − α)λ+ e−λ+ S
(3.17)
with
λ± = K ±
α=
−1
K2 − rτcorr
,
−1 − λ
τcorr
−
,
λ+ − λ−
(3.18)
(3.19)
−1 ]/2. P ( ) above corresponds to choosing
and where K ≡ [r(1 + A) + τcorr
ISI
S
each ISI randomly from a homogeneous Poisson distribution of either rate
λ− (with probability α) or rate λ+ (with probability 1 − α). The inverse relations for r, A, and τcorr as functions of α, λ− , and λ+ are given in appendix A.
To determine I(S; V)/N from equations 3.12 and 3.16, one need only calculate PIVI (V ). For the constant probability synapse, PIVI (V ) can be obtained by the following trick. Because the constant-probability synapse has
no history dependence, vesicle release failures at the constant-probability
synapse are equivalent to failures to receive a spike. Therefore, the vesicle release train produced by a constant-probability synapse with release
probability p receiving input of average rate r is statistically identical to
that produced by a reliable synapse receiving input of average rate pr. This
implies that PIVI (V ) can be determined from PISI (V ) by replacing r in
equations 3.17 through 3.19 by pr.
The difference Iconst /N between I(S; V)/N for the unreliable (p < 1)
and perfectly transmitting (p = 1) synapses is shown in Figure 2. When the
input spike train has no autocorrelations (A = 0; see Figure 2, dashed lines),
the input is a homogeneous Poisson process, and Iconst /N is identical to
that found in section 3.1 (see Figure 1).
When the spike trains have sufficiently strong autocorrelations (see Figure 2, A = 5 and rτcorr ≥ 3), the constant-probability synapse can transmit
more information per vesicle release than a reliable synapse. This reflects
that when inputs are generated by a non-Poisson process, the location of
1146
M. Goldman
'Iconst/N (bits)
1.0
r Wcorr = 0.5
r Wcorr = 3
r Wcorr = 15
0.5
0.0
-0.5
-1.0
-1.5
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Failure probability
Failure probability
Failure probability
A=0
A=1
A=5
Figure 2: Information per vesicle release difference Iconst /N = I(S; V)/N −
Ip=1 (S; V)/N for a constant-probability synapse receiving input generated by
the stationary renewal process of equation 3.17. Columns correspond to inputs
with different decay times of positive autocorrelations (measured relative to the
spike rate r). Individual traces give Iconst /N for the specified autocorrelation
magnitudes A (see the legend). Iconst /N tends to be negative except when the
input is strongly correlated and the transmission probability is very low.
one spike is correlated with the location of other spikes. Observation of a
vesicle release therefore not only identifies the location of the spike that
caused the transmission but also partially identifies the locations of correlated spikes. When the spikes are sufficiently strongly autocorrelated, the
extra information gained about the locations of nontransmitted spikes can
outweigh the noisiness of the constant-probability synapse. In this case,
Iconst /N can be greater than zero. However, the positive information differences tend to be small and occur at only very low transmission probabilities (see Figure 2, <7% transmitted for A = 5, rτcorr = 3; <3% transmitted
for A = 5, rτcorr = 15). Thus, the information per unit time transmitted
by such a synapse would be quite low in this regime (see section 6). The
constant-probability synapse has the same autocorrelation function C(τ ) as
the presynaptic spike train and has maximal noise entropy for a given number of releases. Therefore, positive information differences are attributable
to increases in transmission entropy as the transmission rate becomes lower
rather than to changes in the statistical structure of the releases. For the same
statistical structure, lower transmission rates tend to increase the entropy
per transmission because transmissions are less probable, and thus more
surprising, at lower rates (for a concrete example, consider the low rate
limit of equation 3.4 for the Poisson process). As I will show, positive infor-
Enhancement of Information Transmission Efficiency
1147
mation differences for positively autocorrelated spike trains can be much
more dramatic when the synapse has the dynamics of synaptic depression.
4 Information Transmission for a Stochastic Depressing Synapse
4.1 Model of Synaptic Depression. I next calculate the information
transmitted across a stochastic model synapse with failures of transmission
resulting from post–vesicle release refractoriness. The model is intended to
capture, in an analytically tractable way, both the stochasticity of synaptic
transmission and the essential dynamics of depressing synapses, in which
the average transmission amplitude falls dramatically following a presynaptic spike before recovering over a characteristic timescale (Abbott et al.,
1997; Tsodyks & Markram, 1997; Varela et al., 1997).
I assume a model synapse with a single active zone containing a single vesicle release site. The synapse transmits perfectly when a vesicle is
present at the release site and does not transmit when the vesicle is absent (which I will refer to as being in the depleted state). Following vesicle
release, the synapse becomes refractory until the vesicle is replenished. Vesicle replenishment is governed by a homogeneous Poisson process of rate
1/τD (for similar models, see Vere-Jones, 1966; Goldman, 2000; Matveev &
Wang, 2000). The simplifying assumption that the synapse either transmits
perfectly (when the vesicle is available) or is refractory (when the vesicle
is unavailable) makes the calculation of information transmission for this
model analytically tractable.
The ensemble-averaged behavior of this synapse is that the transmission
probability drops to zero following release and then recovers exponentially
back to one with a time constant τD . Because the synaptic transmission
probability drops all the way to zero following release, this model represents
a very strongly depressing synapse.
Failures of transmission by the model depressing synapse result exclusively from vesicle depletion and are governed by the dynamics of vesicle replenishment. This contrasts with the failures of transmission by the
constant-probability synapse that result from a purely static probability of
vesicle release failure. Studying these two types of synapses allows a comparison between the effects of stochasticity in the presence or absence of
synaptic dynamics.
4.2 Input Generated by a Stationary Renewal Process. In this section,
the calculations of section 3.2 are repeated for the model depressing synapse.
As in that section, general results are first given for an arbitrary stationary
renewal process and then applied to the autocorrelated renewal process
described in section 3.2.2.
4.2.1 General Results. The calculations of H(V)/N and H(V|S)/N for
the depressing synapse follow closely those for the constant-probability
1148
M. Goldman
synapse (section 3.2). This is a consequence of the simplified model of synaptic depression, as described below.
I consider a stationary renewal process with P(s) defined as in equation
3.9. Because the ISIs are independent and the depressing synapse is in an
identical (depleted) state after each transmission, the IVIs are independent
as well. This implies that P(v) takes the form of a product, as in equation 3.10,
and that H(V)/N assumes the form given in equation 3.12. The only difference between H(V)/N for the constant-probability and model-depressing
synapses is in the form of the IVI distribution function PIVI (V ). This can
be found for the depressing synapse by the method of Laplace transforms
(Cox, 1962; Cox & Lewis, 1966; de la Rocha et al., 2002). The calculation is
outlined in appendix B. The resulting distribution is
−1
PIVI (V ) = L
ISI (u + τ −1 )
ISI (u) − P
P
D
.
ISI (u + τ −1 )
1−P
(4.1)
D
The noise entropy per release H(V|S)/n is also of the same form as that
for the constant-probability synapse (see equation 3.15). This is because the
depressing synapse is in an identical (depleted) state not only following
each transmission, but also following each spike. If the synapse is depleted
just before the spike, it remains depleted; if the synapse is not depleted
before the spike, it releases and becomes depleted. As a consequence, the
conditional probabilities P(v|s) factor into a product as in equation 3.13,
leading to H(V|S)/n as in equation 3.15. The difference from the constantprobability synapse is in the expression for the probability Q(b|S ). For the
depressing synapse, Q(b|S ) is given by the exponential probability that
the synapse recovers during the ISI S :
e−S /τD , b = 0
1 − e−S /τD , b = 1.
Q(b|S ) =
(4.2)
H(V|S)/n can be converted to a noise entropy per transmission H(V|S)/N
by dividing by the average probability of vesicle transmission p,
p=
N
=
n
∞
0
dS PISI (S )Q(1|S ),
(4.3)
giving
H(V|S)
=−
N
∞
0
dS PISI (S ) b=0,1 Q(b|S ) log2 (Q(b|S ))
∞
.
0 dS PISI (S )Q(1|S )
(4.4)
The information per vesicle release I(S; V)/N is the difference between
H(V)/N (equation 3.12, using PIVI (V ) from equation 4.1) and H(V|S)/N
(equation 4.4).
Enhancement of Information Transmission Efficiency
'Idep/N (bits)
5
r Wcorr = 0.5
r Wcorr = 3
1149
r Wcorr = 15
4
3
2
1
0
-1
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Failure probability
Failure probability
Failure probability
A=0
A=1
A=5
Figure 3: Information per vesicle release difference Idep /N = I(S; V)/N −
Ip=1 (S; V)/N for the model-depressing synapse receiving input generated by
the stationary renewal process of equation 3.17. Graphs are organized as in Figure 2. Idep /N is positive for a wide range of parameters when the input has
nonzero autocorrelations.
4.2.2 Input with Nonnegative Autocorrelations That Decay Exponentially.
For positively correlated inputs, I(S; V)/N can be much larger for the depressing synapse than for a reliable (p = 1) synapse. I demonstrate this by
evaluating the general results derived above for the exponentially correlated
renewal process (see section 3.2.2, equations 3.17 through 3.19). PIVI (V ) and
p for this process are calculated in appendix B. Here, I give only the resulting
information differences Idep /N = I(S; V)/N − Ip=1 (S; V)/N between the
depressing and reliable synapses.
Figure 3 shows Idep /N as a function of the input parameters and average synaptic failure probability 1 − p. Idep /N > 0 over a wide range
of parameter values, indicating that the depressing synapse can transmit
more information per vesicle release than a reliable synapse. This contrasts
with the corresponding result for the constant-probability synapse (see Figure 2), in which positive information differences occur only for extremely
low transmission probabilities.
For uncorrelated input, the depressing synapse always transmits less information per vesicle release than the perfectly transmitting synapse (Idep <
0; see Figure 3, dashed lines). This is to be expected because for uncorrelated
input, the perfectly transmitting synapse has both maximal total transmission entropy (for a given number of transmissions) and no noise entropy.
However, for correlated input, the perfectly transmitting synapse has nonmaximal total transmission entropy and may transmit less information per
1150
M. Goldman
vesicle release than the depressing synapse (Idep > 0). The positive information differences tend to increase with the magnitude of autocorrelations
A (see Figure 3, thin and thick solid lines). Large increases in Idep occur
around values of p where the correlation and depression timescales match,
τD = τcorr (these values are indicated by the arrows on Figure 4B), and tend
to increase more rapidly for higher values of τcorr (compare the three graphs
in Figure 3). Because higher values of τD correspond to higher failure probabilities (see equation B.13), the rapid increases in Idep shift to higher failure
probabilities when τcorr is increased.
5 Comparison of the Information Transmitted Across the
Constant-Probability and Depressing Synapses
The information differences between the depressing and reliable synapses
reflect differences in both average transmission probability and synaptic
dynamics. To isolate the effects of synaptic dynamics, I next compare the
information transmitted by the depressing synapse to that transmitted by
a constant-probability synapse with the same average transmission probability. The nonnegatively autocorrelated renewal spike train is considered
below. This is followed by a qualitative analysis that enables the information transmission process to be understood in a framework that is more
generally applicable to non-analytically tractable studies.
5.1 Input with Nonnegative Autocorrelations That Decay Exponentially. A comparison of Figures 2 and 3 reveals that the information transmitted by the depressing synapse, Idep (S; V), exceeds that transmitted by the
constant-probability synapse, Iconst (S; V), when the depressing and constantprobability synapses have the same average transmission probability (i.e.,
when p = p so that the average number of transmissions N is the same
for both synapses). Figure 4A shows this explicitly, plotting Idep−const /N =
Idep (S; V)/N − Iconst (S; V)/N as a function of the input parameters and average failure probability. For all parameter values tested, Idep−const /N ≥ 0.
This demonstrates that the depressing synapse transmits more information
about these trains than does the constant probability synapse.
The larger transmission of information per vesicle release by the depressing synapse, compared to the constant probability synapse, primarily
reflects a decrease in the noise entropy H(V|S)/N (see Figure 4C) rather than
an increase in the total entropy H(V)/N (see Figure 4B). This is surprising
in the light of previous information studies (Srinivasan, Laughlin, & Dubs,
1982; Barlow & Foldiak, 1989; Dan et al., 1996; de la Rocha et al., 2002; Goldman et al., 2002; Wang, Liu, Sanchez-Vives, & McCormick, 2003) that have
primarily used decorrelation, or maximization of H(V) for a fixed number
of transmissions, as a measure of information transmission efficiency. For
uncorrelated input, the constant-probability synapse has maximal total entropy per release H(V)/N, reflecting that the transmissions are completely
Enhancement of Information Transmission Efficiency
'Idep-const/N (bits)
A
r Wcorr = 3
r Wcorr = 15
5
4
3
2
1
0
-1
B
'Hdep-const(V)/N
r Wcorr = 0.5
1151
5
4
3
2
1
0
-1
'Hdep-const(V|S)/N
C
1
0
-1
-2
-3
-4
-5
0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0
Failure probability
Correlation
D1
Failure probability
Failure probability
A=0
A=1
A=5
1
0
0
-1
0
W (ms)
-1
500
0
W (ms)
500
Figure 4: Synaptic depression increases the information per vesicle release relative to a constant-probability synapse with the same average transmission probability. (A) Information per vesicle release difference Idep−const /N = Idep /N −
Iconst /N between the depressing and constant-probability synapses. Graphs are
organized as in Figures 2 and 3. (B) The corresponding response entropy difference Hdep−const (V)/N = Hdep (V)/N − Hconst (V)/N is negative for uncorrelated
input (dashed line) but can be positive for correlated input (solid lines). Peak positive values are located near where τD = τcorr (indicated by arrows). These results
reflect the autocorrelation structure of the transmissions (panel D). (C) The noise
entropy difference Hdep−const (V|S)/N = Hdep (V|S)/N − Hconst (V|S)/N is negative for all parameter values and is larger in magnitude than Hdep−const (V)/N.
(D) Autocorrelations C(τ ) of the vesicle releases across a constant-probability
(left) and depressing (right) synapse receiving uncorrelated input. Each synapse
transmitted 50% of the presynaptic spikes. The depressing synapse transmits
more information per vesicle release than the constant-probability synapse even
though the transmissions across the constant-probability synapse are uncorrelated.
1152
M. Goldman
uncorrelated (see Figure 4D, left). By contrast, the depressing synapse has
nonzero autocorrelations (see Figure 4D, right). Thus, for a given failure
probability, the constant-probability synapse has larger total entropy per
release than the depressing synapse in this case (Hdep−const (V)/N < 0; see
Figure 4B, dashed lines). Nevertheless, the depressing synapse transmits
more information per release (Idep−const /N > 0; see Figure 4A, dashed
lines), reflecting that the relative decrease in total entropy for the depressing synapse is more than compensated for by the relative decrease in noise
entropy (H(V|S)/N < 0; see Figure 4C, dashed lines).
The noise entropy per release is always larger for the constant-probability
synapse because a synapse with constant probability of transmission is maximally noisy in the sense that all spikes are transmitted with equal probability. The depressing synapse is less noisy because it has a nonuniform
probability of transmission: spikes that follow long ISIs are transmitted
more reliably, and spikes that follow short ISIs are transmitted less reliably.
For correlated input, the noise entropy differences Hdep−const (V|S)/N are
even larger (see Figure 4C, solid lines). This is because the positively correlated input tends to have clusters of spikes that enhance the nonuniformity
of transmission probabilities across the depressing synapse. The first spike
in a cluster is transmitted with high probability, whereas the latter spikes
are not. Moreover, Hdep−const (V)/N can be greater than zero when the input has positive autocorrelations (see Figure 4B, solid lines). This reflects
the decorrelation of the positively correlated spike trains by the depressing
synapse and, as noted in previous work (Goldman et al., 2002), tends to be
maximal when τD = τcorr (indicated by the arrows in Figure 4B).
5.2 Qualitative Analysis. The previous sections give insight into how
the total entropies compare for the depressing and constant-probability
synapses and how the noise entropies compare. In some cases, such as for
uncorrelated input (see Figure 4, dashed lines), both the total entropy and the
noise entropy are smaller for the depressing synapse than for the constantprobability synapse. To understand why the depressing synapse transmits
more information per vesicle release (Idep−const /N > 0) in these cases
therefore requires the relatively nonintuitive assessment of whether the difference in total entropy, Hdep−const (V)/N, or difference in noise entropy,
Hdep−const (V|S)/N, is larger in magnitude. However, as demonstrated below, it becomes obvious that Idep−const /N > 0 for uncorrelated input if the
information is expressed instead as
I(S; V) = H(S) − H(S|V)
=−
P(s) log2 (P(s))
s
+
v
P(v)
s
P(s|v) log2 (P(s|v)).
(5.1)
(5.2)
Enhancement of Information Transmission Efficiency
1153
When expressed in this manner, I(S; V) gives the amount by which the
uncertainty in the presynaptic spike train ensemble S can be reduced by
observation of the transmissions. The spike train entropy H(S) gives the
uncertainty in the spike train when the transmissions are not known. The
conditional spike train entropy H(S|V) gives the uncertainty remaining after
the transmissions have been observed.
The advantage of writing I(S; V) in this form is that H(S) is independent
of the synaptic properties. Thus, the difference in the information transmitted by two synapses about a given spike train ensemble S is determined
solely by differences in H(S|V), that is, Idep−const (S; V) = −Hdep−const (S|V).
H(S|V) is given
by an average over transmission trains v of the entropy term
h(S|v) = − s P(s|v) log2 (P(s|v)). To understand the qualitative behavior
of H(S|V), I approximate the average over transmission trains by considering a typical set of transmissions (see Figure 5, v). I then compare h(S|v) for
the constant-probability and depressing synapses by estimating the probabilities P(s|v) of various presynaptic spike trains (see Figure 5, s1 , s2 , and
s3 ). More nonuniform distributions of P(s|v) correspond to lower h(S|v) and
therefore to larger I(S; V).
Figure 5A illustrates the information transmission process when the
presynaptic spike train is generated by a homogeneous Poisson process.
Consider a set of transmissions that could be produced by either the constantprobability or depressing synapse (see Figure 5, v). P(s|v) is nonzero only
for presynaptic spike trains with spikes occurring at the times of transmission. I consider three such trains (see Figure 5, s1 , s2 , and s3 ). Because these
trains have equal numbers of spikes, they have equal prior probabilities P(s)
of being generated by the homogeneous Poisson process. For the constantprobability synapse, the posterior probabilities P(s|v) are also equal (see
Figure 5A, top right) because each spike in the presynaptic train has an
equal probability of transmission.
By contrast, these trains have nonuniform probabilities of being transmitted by the depressing synapse (see Figure 5A, bottom right). Spikes
located far away from a previous transmission are more likely to be transmitted by the depressing synapse. Therefore, the spike trains most likely to
be consistent with the shown transmissions are those that do not have any
nontransmitted spikes located far away from the first transmission. This
implies that P(s|v) is lowest for train s3 and greatest for train s1 .
This pictorial argument suggests that P(s|v) is more nonuniform for the
depressing synapse than for the constant-probability synapse, and therefore I(S; V) is greater for the depressing synapse. The nonuniformity of
P(s|v) is not particular to synaptic depression. It would be expected for a
synapse with any form of dynamics. Thus, this result is not particular to
depressing synapses but more generally applies to synapses with dynamic
vesicle release probabilities. Qualitatively, synaptic dynamics allow information to be obtained not only about presynaptic spikes occurring at the
1154
M. Goldman
P(s), or P(s|v) for
const.-prob. synapse
A
s1
s2
WD
B
s1
s2
s3
v
s2
s3
s1
s2
s3
Spike Train
WD
time
P(s|v) with
P(s), or P(s|v) for
depression const.-prob. synapse
v
s1
P(s|v) with
depression
s3
s1
s2
s3
s1
s2
s3
Spike Train
Figure 5: Qualitative comparison of the information transmitted across a depressing and constant-probability synapse with the same average transmission
probabilities. The information transmitted corresponds to the ability to identify which of the many possible presynaptic trains s gives rise to an observed
set of transmissions v. (A) The depressing synapse transmits more information
about a Poisson spike train than does the constant-probability synapse. P(s|v)
is uniform for the constant-probability synapse (top right) but nonuniform for
the depressing synapse (bottom right). This reflects that spikes located far from
the previous transmission are more likely to be transmitted by the depressing synapse. (B) The depressing synapse is particularly well tuned for carrying
information about positively correlated trains. Correlated spike trains tend to
have clusters of spikes (top left) that lead to a strongly nonuniform P(s|v) for
the depressing synapse. In both A and B, we assume that the shown trains are
typical and have equal prior probabilities P(s). In the top right panels, P(s) has
been scaled to be the same height as P(s|v) for the constant-probability synapse.
The arrow labeled τD indicates the time constant of depression. See the text for
details.
Enhancement of Information Transmission Efficiency
1155
times of transmission but also about spikes occurring when there are no
transmissions.
For correlated spike trains, the form of the dynamics is important. Spike
trains generated by a positively correlated process tend to have clustered
spikes. Consider a typical train of transmissions v and three typical presynaptic trains with clustered spikes (see Figure 5B, left). For simplicity, I
assume that these spike trains have equal prior probabilities P(s) and do
not consider less clustered spike trains that would have lower prior probabilities. By the same arguments that were made for uncorrelated input,
the constant-probability synapse has a uniform distribution P(s|v) for these
trains. The depressing synapse, however, has a highly nonuniform distribution because the clustering of spikes makes it relatively likely that trains
will have spikes located either very far away from previous transmissions
(like s1 and s2 ) or very close to the previous transmission (like s3 ). This gives
rise to a much more nonuniform distribution P(s|v) than was found for uncorrelated input (see Figure 5B, left). This suggests why Idep−const /N tends
to be larger for correlated input than for uncorrelated input (see Figure 4A).
The result for correlated input does depend on the form of the dynamics. For example, a facilitating synapse would differ from the depressing
synapse in that it would be more likely, for a fixed number of transmissions,
to transmit multiple spikes from the same cluster and no spikes from other
clusters. Information about the nontransmitted clusters would be lost, and
the representation of the transmitted clusters would be redundant. This
suggests that a facilitating synapse would be less efficient at transmitting
these trains than a depressing synapse.
6 Discussion
I have studied how the information transmission across a synapse depends
on the correlation structure of the presynaptic spike trains, the stochastic nature of synaptic transmission, and the dynamics of synaptic depression. In
summary, when spikes are generated by an uncorrelated process, synaptic
transmission failures reduce the information per vesicle release transmitted
across a synapse. Intuitively, this is because the input has no redundancy that
can be reduced by synaptic failures. For correlated input, synaptic failures
can increase the information transmitted per vesicle release. Information efficiency increases are limited for the constant-probability synapse but can be
large for a depressing synapse. For the nonnegatively autocorrelated inputs
considered here, the information transmitted by the depressing synapse exceeds that transmitted across a constant probability synapse with the same
average transmission probability.
I have used information per vesicle release as the measure of synaptic
transmission efficiency. This measure is easily interpretable and allows a
comparison between synapses that is independent of the precision δT with
which spike times are specified (see section 2). An artifact of the use of this
1156
M. Goldman
measure of efficiency is that in some cases, its maximum value occurs when
the transmission probability equals zero (see Figures 2 and 3). This suggests
that other information-maximization measures might be more suitable in
the low-transmission-probability regime. For example, maximizing information per unit time favors perfect transmission. Maximizing information
per energy usage (Laughlin, de Ruyter van Steveninck, & Anderson, 1998;
Balasubramanian et al., 2001; Levy & Baxter, 2002; Schreiber et al., 2002)
similarly tends to avoid very low transmission probabilities because of the
costs of transmission-independent energy expenditures.
Many studies of information transmission in sensory areas focus on the
role of decorrelation in maximizing information transmission (Srinivasan
et al., 1982; Barlow & Foldiak, 1989; Dan et al., 1996; Goldman et al., 2002;
Wang et al., 2003). Decorrelation corresponds to a maximization of the response entropy that, when the noise entropy is small, also corresponds to an
approximate maximization of the information transmitted. However, the results here show that the noise entropy cannot be neglected. Rather, changes
in the noise entropy are the dominant contributor to the enhanced performance of the depressing synapse over the constant-probability synapse.
The constant-probability synapse is maximally noisy because it transmits
all spikes with equal probability. By contrast, the synaptic failures across the
depressing synapse are much more predictable, with higher probabilities of
failure following short ISIs and lower probabilities following long ISIs.
The noise entropy is likely to be smaller for many synaptic transmission
situations whose behavior is more complex than those modeled here. For
example, noise in the transmission process can be reduced if multiple independently releasing active zones, in either the same or different synapses,
receive the same input (de Ruyter van Steveninck & Laughlin, 1996; Manwani & Koch, 2000). In such cases, maximization of the total transmission
entropy (see Figure 4B) would assume a relatively larger role in maximizing
the information transmission. In the limit that the total transmission entropy
dominates, the result of maximizing the information per transmission for a
given fraction of transmissions would be expected to approach that found
in previous decorrelation studies (Goldman et al., 2002). However, the dominance of the noise entropy in the examples studied here suggests that for
many situations, the noise entropy is a significant factor that cannot be neglected. This is consistent with studies of other adaptation (Wainwright,
1999; Brenner, Bialek, & de Ruyter van Stevenick, 2000; Fairhall, Lewen,
Bialek, & de Ruyter van Steveninck, 2001) or refractory (Berry & Meister,
1998) processes that implicate a significant role for noise reduction in the
information transmission process.
Direct numerical calculations of information (de Ruyter van Steveninck
et al., 1997; Strong et al., 1998) explicitly account for noise entropies but are
limited in practice by the need to use either large time bins or brief duration word sizes. These constraints can be problematic when considering
spike trains that may have high temporal precision but long timescales of
Enhancement of Information Transmission Efficiency
1157
autocorrelations. I handle this problem by analytically solving the synaptic information transmission problem for simplified models of positively
correlated spike trains and depressing synapses. Attempts to include more
realistic features of spike trains and synapses in calculation of the Shannon
information have worked in limits that consider only total transmission
entropies (Goldman et al., 2002; de la Rocha et al., 2002), or used approximate but more easily calculated measures of the information transmission
(Fuhrmann et al., 2002).
The ideal measure of information transmission efficacy must consider the
synapse in the context of the whole neuron or network. Calculations of the
information that a postsynaptic neuron’s spike train conveys about a particular presynaptic input have thus far been limited to constant-transmissionprobability models of stochastic synapses (Zador, 1998; Manwani & Koch,
2001) or have ignored synaptic stochasticity altogether (London, Schreibman, Hausser, Larkum, & Segev, 2002). An interesting open question is
how synaptic dynamics and stochasticity may be incorporated into such
models of information transmission by the postsynaptic neuron.
Because the calculations here and elsewhere are applicable only to limited
situations, qualitative analysis provides a valuable tool for gaining insight
into the synaptic transmission process. The framework presented here formulates the information transmission process as a stimulus reconstruction
problem in which the likelihoods of different presynaptic spike trains are
estimated based on a given set of transmissions. This analysis, for example,
shows clearly how synaptic dynamics can increase the information transmitted across a synapse by providing information about the locations of
spikes that are not themselves transmitted (see Figure 5). For uncorrelated
input (see Figure 5A), the presence of synaptic dynamics, rather than the
exact form of the dynamics, was seen to be the essential feature that leads to
increases in information transmission. For correlated input, the dynamics
of synaptic depression were seen to be particularly well adapted to transmitting information about positively correlated inputs. In both examples,
general insights were drawn without requiring a detailed mathematical
model of the synapse. I expect that this framework will provide similar insights for systems with different spike train autocorrelations and different
synaptic dynamics.
Appendix A: Autocorrelation Function of a Stationary Renewal Process
Spike Train
Here, the normalized autocorrelation function C(τ ) of a stationary renewal
process spike train is related to the ISI distribution PISI (S ) from which the
spike train was generated. The normalized autocorrelation function C(τ )
1158
M. Goldman
for a point-process spike train s(t) is defined by
C(τ ) ≡
s(t)s(t + τ ) − r2
,
r2
(A.1)
where (·) denotes ensemble averaging. Following the notation of Cox and
Lewis (1966), C(τ ) can be expressed as
C(τ ) =
m f (τ ) − r
,
r
(A.2)
where m f (τ ) gives the probability density that some pair of events is separated by a time interval τ . m f (τ ) is expressed as a sum of probabilities fi (τ )
that the ith ISI is in the interval (τ, τ + dτ ),
m f (τ ) =
∞
fi (τ ).
(A.3)
i=1
For a renewal process, fi (τ ) is given by convolving the ISI distribution
PISI (S ) with itself i times. Laplace transforming the above equation thus
gives
f (u) =
m
∞
ISI (u)]i =
[P
i=1
ISI (u)
P
.
ISI (u)
1−P
(A.4)
Inverse Laplace-transforming and substituting into equation A.2 gives C(τ ).
For the ISI distribution given in equation 3.17 with Laplace transform
given by equation B.4, substitution into equation A.4 gives exponentially
decaying autocorrelations of the form C(τ ) = Ae−τ/τcorr , with A, τcorr , and
the average rate r given by
A=
α(1 − α)(λ− − λ+ )2
λ− λ+
(A.5)
1
(1 − α)λ− + αλ+
(A.6)
τcorr =
r=
λ− λ+
.
(1 − α)λ− + αλ+
(A.7)
It can be shown that this ISI distribution can assume any arbitrary coefficient of variation between one and ∞, depending on the choice of the
distribution parameters (Cox, 1962).
Enhancement of Information Transmission Efficiency
1159
Appendix B: Calculation of PIVI (V ) and p for the Stationary Renewal
Spike Train
Below, the IVI distributions PIVI (V ) are derived for the constant-probability
or model depressing synapse receiving input generated by a stationary renewal process.
For the constant-probability synapse, PIVI (V ) is obtained by noting that
the probability of having an IVI of duration V can be broken into two terms
as
PIVI (V ) = pPISI (V )
+ (1 − p)
0
(B.1)
V
dS PISI (S )PIVI (V − S ).
(B.2)
The first term gives the probability that the first ISI is of duration V and
the vesicle releases. The second term gives the sum over all possible ways
that an IVI of duration V could occur when the first ISI does not cause a
vesicle release.
PIVI (V ) is obtained from the Laplace transform of the above equation:
IVI (u) =
P
ISI (u)
pP
.
ISI (u)
1 − (1 − p)P
(B.3)
For the exponentially correlated spike train model described by equation
3.17, the Laplace transform of PISI (S ) is
ISI (u) = αλ− /(u + λ− ) + (1 − α)λ+ /(u + λ+ ).
P
(B.4)
Inserting the above into equation B.3 and inverse Laplace transforming
gives that PIVI (V ) is identical in form to the ISI distribution PISI (V ) (see
equation 3.17) except for the replacement of the average rate r by pr. This
result reflects that failures to transmit and failures to receive a presynaptic
spike are equivalent for the constant-probability synapse.
For the model-depressing synapse, PIVI (V ) can similarly be decomposed into two terms corresponding to an IVI of duration V resulting
from the first ISI or multiple ISIs:
PIVI (V ) = PISI (V )(1 − e−V /τD )
V
+
dS PISI (S )e−S /τD PIVI (V − S ).
(B.5)
(B.6)
0
Solving this equation by using Laplace transforms gives
−1
IVI (u) = PISI (u) − PISI (u + τD ) .
P
ISI (u + τ −1 )
1−P
D
(B.7)
1160
M. Goldman
For the exponentially correlated spike train, PIVI (V ) is obtained by substituting equation B.4 into equation B.7 and inverse Laplace transforming.
The resulting expression gives PIVI (V ) as a linear combination of four
exponential distributions with inverse time constants λ1 = λ− , λ2 = λ+ ,
−1 + τ −1 and corresponding weights γ :
λ3 = τD−1 , and λ4 = τcorr
i
D
PIVI (V ) =
4
γi λi e−λi V ,
(B.8)
i=1
where
γ1 =
αλ3 (λ2 − λ1 + λ3 )
(λ4 − λ1 )(λ3 − λ1 )
γ2 =
(1 − α)λ3 (λ1 − λ2 + λ3 )
(λ4 − λ2 )(λ3 − λ2 )
(B.10)
γ3 =
−1 − λ )
r(τcorr
3
(λ3 − λ2 )(λ3 − λ1 )
(B.11)
γ4 =
−Arλ3
.
(λ4 − λ2 )(λ4 − λ1 )
(B.12)
(B.9)
The weights γi can be negative so, unlike α in the ISI distribution (see equation 3.17), they cannot be interpreted as probabilities.
The average transmission probabilities p = N/n are obtained by substituting PISI (S ) (see equation 3.17) into equation 4.3. The result is:
p=1−
αλ1
(1 − α)λ2
−
.
λ1 + λ3
λ2 + λ3
(B.13)
For uncorrelated input (α = 0 or 1), p depends on only the product rτD . For
correlated input, p depends more generally on the form of the autocorrelations. This implies that for a given value of p, rτD differs across the graphs
shown in Figures 3 and 4. The arrows labeling τD = τcorr in Figure 4B give
an example of this dependence.
Acknowledgments
This work was supported by NIH grants MH58754 and MH60651, NSF grant
IBN-9817194, the Sloan Center for Theoretical Neurobiology at Brandeis
University, the W.M. Keck Foundation, and the Howard Hughes Medical
Institute. I thank Larry Abbott and Dan Butts for helpful discussions and
comments on the manuscript and Xiaohui Xie for helpful comments on the
manuscript.
Enhancement of Information Transmission Efficiency
1161
References
Abbott, L. F., Varela, J. A., Sen, K., Nelson, S. B. (1997). Synaptic depression and
cortical gain control. Science, 275, 220–223.
Attneave, F. (1954). Some informational aspects of visual perception. Psych. Rev.,
61, 183–193.
Baddeley, R., Abbott, L. F., Booth, M. C., Sengpiel, F., Freeman, T., Wakeman,
E. A., & Rolls, E. T. (1997). Responses of neurons in primary and inferior
temporal visual cortices to natural scenes. Proc. R. Soc. Lond. B Biol. Sci., 264,
1775–1783.
Balasubramanian, V., Kimber, D., Berry, M. J. II. (2001). Metabolically efficient
information processing. Neural Comput., 13, 799–815.
Barlow, H. B. (1961). Possible principles underlying the transformation of sensory messages. In W. A. Rosenblith (Ed.), Sensory communication (pp. 217–234).
Cambridge, MA: MIT Press.
Barlow, H. B., & Foldiak, P. (1989). Adaptation and decorrelation in the cortex. In
R. Durbin, C. Miall, & G. Mitchinson (Eds.), The computing neuron. Reading,
MA: Addison-Wesley.
Berry, M. J., & Meister, M. (1998). Refractoriness and neural precision. J. Neurosci.,
18, 2200–2211.
Brenner, N., Bialek, W., & de Ruyter van Steveninck, R. (2000). Adaptive rescaling
maximizes information transmission. Neuron, 26, 695–702.
Brenner, N., Strong, S. P., Koberle, R., Bialek, W., & de Ruyter van Steveninck,
R. R. (2000). Synergy in a neural code. Neural Comput., 12, 1531–1552.
Cox, D. R. (1962). Renewal theory. London: Methuen.
Cox, D. R., & Lewis, P. A. W. (1966). The statistical analysis of series of events.
London: Methuen.
Dan, Y., Atick, J. J., & Reid, R. C. (1996). Efficient coding of natural scenes in
the lateral geniculate nucleus: Experimental test of a computational theory.
J. Neurosci., 16, 3351–3362.
de Ruyter van Steveninck, R. R., & Laughlin, S. B. (1996). The rate of information
transfer at graded-potential synapses. Nature, 379, 642–645.
de Ruyter van Steveninck, R. R., Lewen, G. D., Strong, S. P., Koberle, R., & Bialek,
W. (1997). Reproducibility and variability in neural spike trains. Science, 275,
1805–1808.
de la Rocha, J., Nevado, A., & Parga, N. (2002). Information transmission by
stochastic synapses with short-term depression: Neural coding and optimization. Neurocomputing, 44–46, 85–90.
Fairhall, A. L., Lewen, G. D., Bialek, W., & de Ruyter van Steveninck, R. R. (2001).
Efficiency and ambiguity in an adaptive neural code. Nature, 412, 787–792.
Fuhrmann, G., Segev, I., Markram, H., & Tsodyks, M. (2002). Coding of temporal
information by activity-dependent synapses. J. Neurophysiol., 87, 140–148.
Goldman, M. S. (2000). Computational implications of activity-dependent neuronal
processes. Unpublished doctoral dissertation, Harvard University.
Goldman, M. S., Maldonado, P., & Abbott, L. F. (2002). Redundancy reduction
and sustained firing with stochastic depressing synapses. J. Neurosci., 22,
584–591.
1162
M. Goldman
Laughlin, S. B. (2001). Energy as a constraint on the coding and processing of
sensory information. Curr. Opin. Neurobiol., 11, 475–480.
Laughlin, S. B., de Ruyter van Steveninck, R. R., & Anderson, J. C. (1998). The
metabolic cost of neural information. Nat. Neurosci., 1, 36–41.
Levy, W. B., & Baxter, R. A. (2002). Energy-efficient neuronal computation via
quantal synaptic failures. J. Neurosci., 22, 4746–4755.
London, M., Schreibman, A., Hausser, M., Larkum, M. E., & Segev, I. (2002). The
information efficacy of a synapse. Nat. Neurosci., 5, 332–340.
Maass, W., & Zador, A. M. (1999). Dynamic stochastic synapses as computational
units. Neural Comput., 11, 903–917.
Manwani, A., & Koch, C. (2001). Detecting and estimating signals over noisy
and unreliable synapses: Information theoretic analysis. Neural Comput., 13,
1–33.
Matveev, V., & Wang, X.-J. (2000). Differential short-term synaptic plasticity and
transmission of complex spike trains: To depress or to facilitate? Cereb. Cortex,
10, 1143–1153.
Murthy, V. N., Sejnowski, T. J., & Stevens, C. F. (1997). Heterogeneous release
properties of visualized individual hippocampal synapses. Neuron, 18, 599–
612.
Natschlager, T., & Maass, W. (2001). Computing the optimally fitted spike train
for a synapse. Neural Comput., 13, 2477–2494.
Schreiber, S., Machens, C. K., Herz, A. V. M., & Laughlin, S. B. (2002). Energyefficient coding with discrete stochastic events. Neural Comput., 14, 1323–1346.
Shannon, C., & Weaver, W. (1949). The mathematical theory of communication. Urbana: University of Illinois Press.
Srinivasan, M. V., Laughlin, S. B., & Dubs, A. (1982). Predictive coding: A fresh
view of inhibition in the retina. Proc. R. Soc. Lond. B. Biol. Sci., 216, 427–459.
Stevens, C. F., & Wang, Y. (1995). Facilitation and depression at single central
synapses. Neuron, 14, 795–802.
Strong, S. P., Koberle, R., de Ruyter van Steveninck, R. R., & Bialek, W. (1998).
Entropy and information in neural spike trains. Phys. Rev. Lett., 80, 197–200.
Tsodyks, M. V., & Markram, H. (1997). The neural code between neocortical
pyramidal neurons depends on neurotransmitter release probability. Proc.
Natl. Acad. Sci. U.S.A., 94, 719–723.
Varela, J., Sen, K., Gibson, J., Fost, J., Abbott, L. F., & Nelson, S. B. (1997). A
quantitative description of short-term plasticity at excitatory synapses in
layer 2/3 of rat primary visual cortex. J. Neurosci., 17, 7926–7940.
Vere-Jones, D. (1966). Simple stochastic models for the release of quanta of transmitter from a nerve terminal. Aust. J. Stat., 8, 53–63.
Wainwright, M. J. (1999). Visual adaptation as optimal information transmission.
Vision Res., 39, 3960–3974.
Wang, X. J., Liu, Y., Sanchez-Vives, M. V., & McCormick, D. A. (2003). Adaptation
and temporal decorrelation by single neurons in the primary visual cortex.
J. Neurophysiol., 89, 3279–3293.
Zador, A. (1998). Impact of synaptic unreliability on the information transmitted
by spiking neurons. J. Neurophysiol., 79, 1219–1229.
Received May 15, 2003; accepted November 6, 2003.