Applications of free probability and random matrix

CMA 2007
Applications of free probability and random matrix
theory
Øyvind Ryan
January 2008
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Some important concepts from classical probability
I
Random variables are functions (i.e. they commute w.r.t.
multiplication) with a given p.d.f. (denoted f )
I
Expectation (denoted E ) is integration
I
Independence
I
Additive convolution (∗) and the logarithm of the Fourier
transform
I
Multiplicative convolution
I
Central limit law, with special role of the Gaussian law
¢
¢∗n
¡¡
Poisson distribution Pc : The limit of 1 − nc δ(0) + nc δ(1)
as n → ∞.
I
I
Divisibility: For a given a, nd i.i.d. b1 , ..., bn such that
fa = fb1 +···+bn .
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
I
Can we nd a more general theory, where the random variables
are matrices (or more generally, operators), with their
eigenvalue distribution (or spectrum) taking the role as the
p.d.f.?
I
What are the analogues to the above mentioned concepts for
this theory? In particular, what plays the role as central limit?
I
What are the applications of such a theory?
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Free probability
Free probability was developed as a probability theory for random
variables which do not commute, like matrices
I The random variables are elements in a unital ∗-algebra
(denoted A), typically B (H), or Mn (C).
I Expectation (denoted φ) is a normalized linear functional on A.
The pair (A, φ) is called a noncommutative probability space.
I For matrices, φ will be the normalized trace trn , dened by
n
1X
trn (a) =
aii .
n
i =1
I
For random matrices, we set φ(a) = τn (a) = E (trn (a)) is
dened by.
Useful way to think about free probability: Two independent
random matrices, where the eigenvectors for one of them point
in each direction with equal probability, then the two matrices
are "free" (to be dened).
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The full circle law
Let Xn = √1n Yn where Yn is n × n and has i.i.d. complex standard
Gaussian entries. This is a "central limit candidate". What
happens when n is large? The eigenvalues converge to what is
called the full circle law. Here for n = 500.
1
0.8
0.6
0.4
0.2
0
−0.2
−0.4
−0.6
−0.8
−1
−1
−0.5
0
0.5
1
plot(eig( (1/sqrt(1000)) * (randn(500,500) +
j*randn(500,500)) ),'kx')
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The semicircle law
35
30
25
20
15
10
5
0
−3
−2
−1
0
1
2
3
A = (1/sqrt(2000)) * (randn(1000,1000) +
j*randn(1000,1000));
A = (sqrt(2)/2)*(A+A');
hist(eig(A),40)
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The Marchenko Pastur law
What happens with the eigenvalues of N1 Xn XH
n when Xn is an
n × N random matrix with standard complex Gaussian entries?
I One can show that
µµ
¶p ¶
X
X ak
1
H
Xn Xn
=
c l (π̂)−1 +
τn
.
N
N 2k
π̂∈NC2p
I
k
Convergence is "almost sure", i.e. we have very accurate
eigenvalue prediction when the matrices are large.
When Nn → c, the eigenvalue distribution converges to the
Marchenko Pastur law with parameter c, denoted µ Nn , with
density
p
(x − a)+ (b − x )+
1 +
µc
f (x ) = (1 − ) δ(x ) +
,
(1)
c
2π cx
√
√
where (z )+ = max(0, z ), a = (1 − c )2 and b = (1 + c )2 .
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Four dierent Marchenko Pastur laws µ Nn are drawn.
1.6
c=0.5
c=0.2
c=0.1
c=0.05
1.4
1.2
Density
1
0.8
0.6
0.4
0.2
0
0.5
1
Øyvind Ryan
1.5
2
2.5
3
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Motivation for free probability
One can show that for the Gaussian random matrices we
considered, the limits
¡
¢
¡
¢
φ X i1 Y j1 · · · X il Y jl = lim trn Xin1 Ynj1 · · · Xinl Ynjl
n→∞
exist. If we linearly extend the linear functional φ to all polynomials
in A and B, the following can be shown:
Theorem
If Pi , Qi are polynomials in X and Y respectively, with 1 ≤ i ≤ l ,
and φ(Pi (X )) = 0, φ(Qi (Y )) = 0 for all i, then
φ (P1 (X )Q1 (Y ) · · · Pl (X )Ql (Y )) = 0.
This motivates the denition of freeness, which is the analogy to
independence.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Denition of freeness
Denition
A family of unital ∗-subalgebras (Ai )i ∈I is called a free family if


aj ∈ Aij


⇒ φ(a1 · · · an ) = 0. (2)
i1 6= i2 , i2 6= i3 , · · · , in−1 6= in


φ(a1 ) = φ(a2 ) = · · · = φ(an ) = 0
A family of random variables ai is called a free family if the algebras
they generate form a free family.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
The free central limit theorem
Theorem
If
I
a1 , ..., an are free and self-adjoint,
I
φ(ai ) = 0,
I
I
φ(ai2 = 1,
supi |φ(aik )| < ∞ for all k,
√
then the sequence (a1 + · · · + an )/ n converges in distribution to
the semicircle law.
In free probability, the semicircle law thus
√ has the role of the
1
Gaussian law. it's density is density 2π 4 − x 2
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Similarities between classical and free probability
1. Additive convolution ¢: The p.d.f. of the sum of free random
variables. The role of the logarithm of the Fourier transform is
now taken by the R-transform, which satises
Rµa ¢µb (z ) = Rµa (z ) + Rµb (z ).
2. The S-transform: Transform on probability distributions which
satises Sµa £µb (z ) = Sµa (z )Sµb (z )
3. Poisson distributions have their analogy in the free Poisson
distributions: These are given by the Marcenko Pastur laws µc
with parameter c, which also can be written as the limit of
¡¡
¢
¢¢n
1 − nc δ(0) + nc δ(1)
as n → ∞
4. Innite divisibility: There exists an analogy to the Lévy-Hin£in
formula for innite divisibility in classical probability.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Main usage of free probability in my papers
I
Let X and Y be random matrices. How can we make a good
prediction of the eigenvalue distribution of X when one has the
eigenvalue distribution of XY and Y? Simplest case is when
one assumes that Y is Gaussian (Noise). What about the
eigenvectors?
I
Assume that we have the eigenvalue distribution of
1
H
N (R + X)(R + X) , where R and X are independent n × N
random matrices, with X Gaussian. If the columns of R are
realizations of some random vector r, what is the covariance
matrix E (ri rj∗ )?
I
Multiplicative free convolution with the Marchenko Pastur law
is particularly useful. This has an ecient implementation.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Application of free probability
The capacity per receiving antenna of a channel with n × m
channel matrix H and signal to noise ratio ρ = σ12 is given by
µ
¶
n
1
1
1X
1
H
C = log2 det In +
HH
=
log2 (1 + 2 λl )
n
mσ 2
n
σ
(3)
l =1
where λl are the eigenvalues of m1 HHH . We would like to estimate
C.
To estimate C , we will use free probability tools to estimate the
eigenvalues of m1 HHH based on some observations Ĥi
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Observation model
The following is a much used observation model:
Ĥi = H + σ Xi
(4)
where
I
The matrices are n × m (n is the number of receiving
antennas, m is the number of transmitting antennas)
I
Ĥi is the measured MIMO matrix,
Xi is the noise matrix with i.i.d standard complex Gaussian
I
entries.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Existing ways to estimate the channel capacity
Several channel capacity estimators have been used in the literature:
³
´
1 PL
1
H
Ĥ
Ĥ
C1 = nL
log
det
I
+
n
i =1
mσ 2 i i ´
³2
1
1 PL
C2 = n log2 det In + Lσ2 m i =1 Ĥi ĤH
(5)
i
³
´
P
P
C3 = n1 log2 det In + σ21m ( L1 Li=1 Ĥi )( L1 Li=1 Ĥi )H )
Why not try to formulate an estimator based on free probability
instead?
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
2.6
2.4
2.2
Capacity
2
1.8
1.6
1.4
1.2
1
True capacity
C1
C2
C3
0.8
0
5
10
15
20
Number of observations
25
30
Comparison of the classical capacity estimators for various number
of observations. σ 2 = 0.1, n = 10 receive antennas, m = 10
transmit antennas. The rank of H was 3.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Main free probability result we will use
Dene
Γn =
Wn =
1
Rn RHn
N
1
(Rn + σ Xn )(Rn + σ Xn )H ,
N
where Rn and Xn are independent n × N random matrices, Xn is
complex, standard, Gaussian.
Theorem
If e .e .d .(Γn ) → νΓ , then e .e .d .(Wn ) → νW where νW is uniquely
identied by
νW rµc = (νΓ rµc ) ¢ δσ2
(r = "the opposite of £").
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Realization of the theorem for the problem at hand
Form the compound observation matrix
σ
Ĥ1...L = H1...L + √ X1...L , where
L
i
1 h
Ĥ1...L = √ Ĥ1 , Ĥ2 , ..., ĤL ,
L
1
H1...L = √ [H, H, ..., H] ,
L
X1...L = [X1 , X2 , ..., XL ] .
For the problem at hand, the theorem takes the form
³
´
n ≈ ν1
n
ν 1 Ĥ1...L ĤH rµ mL
¢ δσ 2
rµ
H
H1...L H
mL
m
m
1...L
1...L
(6)
1
H
Since m1 H1...L HH
1...L = m HH , we can now estimate the moments
of m1 HHH from the moments of the observation matrix
1
H
m Ĥ1...L Ĥ1...L , and thereby estimate the eigenvalues, and hence the
channel capacity.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
Free probability based estimator for the moments of the
channel matrix
Can also be written in the following way for the rst four moments:
ĥ1 = h1 + σ 2
ĥ2 = h2 + 2σ 2 (1 + c )h1 + σ 4 (1 + c )
ĥ3 = h3 + 3¡σ 2 (1 + c )h2 ¢+ 3σ 2 ch12
+3σ 4¡ c 2 + 3c + 1¢ h1
+σ 6 c 2 + 3c + 1
ĥ4 = h4 + 4σ 2 (1 + c )h3 + 8σ 2 ch2 h1
+σ 4 (6c 2 + 16c + 6)h2
+14σ 4 c (1 + c )h12
+4σ 6¡(c 3 + 6c 2 + 6c + 1¢)h1
+σ 8 c 3 + 6c 2 + 6c + 1 ,
where ĥi are the moments of the observation matrix
hi are the moments of m1 HHH .
Øyvind Ryan
(7)
1
H
m Ĥ1...L Ĥ1...L ,
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
2.6
2.4
2.2
Capacity
2
1.8
1.6
1.4
1.2
1
True capacity
C
f
Cu
0.8
0
5
10
15
20
Number of observations
25
30
Comparison of Cf and Cu for various number of observations.
σ 2 = 0.1, n = 10 receive antennas, m = 10 transmit antennas. The
rank of H was 3.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
4.5
4
Capacity
3.5
True capacity, rank 3
C , rank 3
f
True capacity, rank 5
C , rank 5
3
f
True capacity, rank 6
Cf, rank 6
2.5
2
1.5
0
10
20
30
40
50
60
Number of observations
70
80
90
100
Cf for various number of observations. No phase o-set/phase
drift. σ 2 = 0.1, n = 10 receive antennas, m = 10 transmit
antennas. The rank of H was 3, 5 and 6.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
List of papers
I
Free Deconvolution for Signal Processing Applications.
Submitted to IEEE Trans. Inform. Theory.
arxiv.org/cs.IT/0701025.
I
Multiplicative free Convolution and Information-Plus-Noise
Type Matrices. arxiv.org/math.PR/0702342.
I
Channel Capacity Estimation using Free Probability Theory.
Submitted to IEEE. Trans. Signal Process.
arxiv.org/abs/0707.3095.
I
Random Vandermonde Matrices-Part I: Fundamental results.
Work in progress.
I
Random Vandermonde Matrices-Part II: Applications to
wireless applications. Work in progress.
I
Vandermonde Frequency Division Multiplexing: An Approach
for Cognitive Radio. Work in progess.
Øyvind Ryan
Applications of free probability and random matrix theory
CMA 2007
Applications of free probability and random matrix theory
This talk is available at
http://heim.i.uio.no/∼oyvindry/talks.shtml.
My publications are listed at
http://heim.i.uio.no/∼oyvindry/publications.shtml
Øyvind Ryan
Applications of free probability and random matrix theory