Introduction to Quantization - Faculty of Physics University of Warsaw

Introduction to Quantization
Jan Dereziński
Dept. of Math. Methods in Phys.,
Faculty of Physics, University of Warsaw
ul. Pasteura 5, 02-093 Warszawa, Poland
email [email protected]
March 7, 2017
1
Introduction
1.1
Basic classical mechanics
Basic classical mechanics takes place in phase space Rd ⊕ Rd . The variables are the positions
xi , i = 1, . . . , d, and the momenta pi , i = 1, . . . , d. Real-valued functions on Rd ⊕ Rd are
called observables. (For example, positions and momenta are observables). The space of
observables is equipped with the (commutative) product bc and with the Poisson bracket
{b, c} = ∂xi b∂pi c − ∂pi b∂xi c .
(We use the summation convention of summing wrt repeated indices). Thus in particular
{xi , xj } = {pi , pj } = 0,
{xi , pj } = δji .
(1.1)
The dynamics is given by a real function on Rd ⊕ Rd called the (classical) Hamiltonian
H(x, p). The equations of motion are
dx(t)
dt
dp(t)
dt
=
∂p H(x(t), p(t)),
=
−∂x H(x(t), p(t)).
We treat x(t), p(t) as the functions of the initial conditions
x(0) = x,
p(0) = p.
More generally, the evolution of an observable b(x, p) is given by
d
b(x(t), p(t)) = {b, H}(x(t), p(t)).
dt
1
The dynamics preserves the product (this is obvious) and the Poisson bracket:
bc(x(t), p(t)) = b(x(t), p(t))c(x(t), p(t)),
{b, c}(x(t), p(t)) = {b(x(t), p(t)), c(x(t), p(t))},
Examples of classical Hamitonians:
particle in electrostatic and
1
(p − A(x))2 + V (x),
2m
1 ij
pi g (x)pj ,
2
1 2 ω2 2
p +
x ,
2
2
1
1
(p1 − Bx2 )2 + (p2 + Bx1 )2 ,
2
2
1 ij
1
a pi pj + bij xi pj + cij xi xj .
2
2
magnetic potentials
particle in curved space
harmonic oscillator
particle in constant magnetic field
general quadratic Hamiltonian
1.2
Basic quantum mechanics
Let ~ be a positive parameter, typically small.
Basic quantum mechanics takes place in the Hilbert space L2 (Rd ). Self-adjoint operators
on L2 (Rd ) are called observables. For a pair of such operators A, B we have their product
AB. (Note that we disregard the issues that arise with unbounded operators for which the
product is problematic). From the product one can derive their commutative Jordan product
1
2 (AB + BA) and their commutator [A, B]. The dynamics is given by a self-adjoint operator
H called the Hamiltonian. On the level of the Hilbert space the evolution equation is
i~
dΨ
= HΨ(t), Ψ(0) = Ψ,
dt
it
so that Ψ(t) = e ~ H Ψ. On the level of observables,
~
it
dA(t)
= i[H, A(t)], A(t0 ) = A,
dt
it
so that A(t) = e ~ H Ae− ~ H . The dynamics preserves the product:
it
it
it
it
it
it
e ~ H ABe− ~ H = e ~ H Ae− ~ H e ~ H Be− ~ H .
We have distinguished observables: the positions x̂i , i = 1, . . . , n, and the momenta p̂i :=
~
i
i ∂x , i = 1, . . . , n. They satisfy
[x̂i , x̂j ] = [p̂i , p̂j ] = 0,
2
[x̂i , p̂j ] = i~δji .
Examples of quantum Hamiltonians
particle in electrostatic and
1
(p̂ − A(x̂))2 + V (x̂),
2m
1 −1/4
g
(x̂)p̂i g ij (x̂)g 1/2 (x̂)p̂j g −1/4 (x̂),
2
1 2 ω2 2
p̂ +
x̂ ,
2
2
1
1
(p̂1 − B x̂2 )2 + (p̂2 + B x̂1 )2 ,
2
2
1 ij
1
i
a p̂i p̂j + bj x̂i p̂j + cij x̂i x̂j .
2
2
magnetic potentials
particle in curved space
harmonic oscillator
particle in constant magnetic field
general quadratic Hamiltonian
1.3
Concept of quantization
Quantization usually means a linear transformation with good properties, which to a function
on phase space b : R2d → C associates an operator Op• (b) acting on the Hilbert space
L2 (Rd ). (The superscript • stands for a possible decoration indicating the type of a given
quantization).
(Sometimes we will write Op• (b(x, p)) for Op• (b) – x, p play the role of coordinate
functions on the phase space and not its concrete points).
Here are desirable properties of a quantization:
(1) Op• (1) = 1l, Op• (xi ) = x̂i , Op• (pj ) = p̂j .
(2) 21 Op• (b)Op• (c) + Op• (c)Op• (b) ≈ Op• (bc).
(3) [ Op• (b), Op• (c)] ≈ i~Op• ({b, c}).
(4) [ Op• (b), Op• (c)] = i~Op• ({b, c}) if b is a 1st order polynomial.
The function b will be called the symbol (or dequantization) of the operator B.
Note that (1) implies that (3) and (4) is true at least if b is a 1st order polynomial.
1.4
The role of the Planck constant
Recall that the position operator x̂i , is the multiplication by xi and the momentum operator
is p̂i := ~i ∂xi . Thus we treat the position x̂ as the distinguished physical observable. The
momentum is scaled. This is the usual convention in physics.
Let Op• (b) stand for the quantization with ~ = 1. The quantization with any ~ will be
denoted by Op•~ (b). Note that we have the relationship
Op•~ (b) = Op• (b~ ),
b~ (x, p) = b(x, ~p).
However this convention breaks the symplectic invariance of the phase space. In some
situations it is more natural to use the Planck constant
√ differently and to use the position
operator
x̃
which
is
the
multiplication
operator
by
~xi , and the momentum operator
i
√
~
p̃ := i ∂xi . Note that they satisfy the usual commutation relations
[x̃i , p̃j ] = i~δij .
3
The corresponding quantization of a function b is
•
f ~ (b) := Op• (b̃~ ),
Op
so that
√
√
b̃~ (x, p) = b( ~x, ~p).
•
(1.2)
•
f ~ (xi ) = x̃i ,
Op
f ~ (pi ) = p̃i .
Op
The advantage of (1.2) is that positions and momenta are treated on equal footing. This
approach is typical when we consider coherent states.
Of course, both approaches are unitary equivalent. Indeed, introduce the scaling
τλ Φ(x) = λ−d/2 Φ λ−1/2 x .
Then
•
f ~ (b~ ) = τ~1/2 Op• (b~ )τ~−1/2 .
Op
1.5
Aspects of quantization
Quantization has many aspects in contemporary mathematics and physics.
1. Fundamental formalism
– used to define a quantum theory from a classical theory;
– underlying the emergence of classical physics from quantum physics (Weyl-WignerMoyal, Wentzel-Kramers-Brillouin).
2. Technical parametrization
– of operators used in PDE’s (Maslov, 4 volumes of Hörmander);
– of observables in quantum optics (Nobel prize for Glauber);
– signal encoding.
3. Subject of mathematical research
– geometric quantization;
– deformation quantization (Fields medal for Kontsevich!);
4. Harmonic analysis
– on the Heisenberg group;
– special approach for more general Lie groups and symmetric spaces.
We will not discuss (3), where the starting point is a symplectic or even a Poison manifold.
We will concentrate on (1) and (2), where the starting point is a (linear) symplectic space,
or sometimes a cotangent bundle.
A seperate subject is quantization of systems with an infinite number of degrees of
freedom, as in QFT, where it is even nontrivial to quantize linear dynamics.
4
2
Preliminary concepts
2.1
Fourier tranformation
We adopt the following definition of the Fourier transform.
Z
Ff (ξ) := e−iξx f (x)dx.
The inverse Fourier transform is given by
F
−1
−d
Z
g(x) = (2π)
eixξ g(ξ)dξ.
Formally, F −1 F = 1l can be expressed as
Z
(2π)−d ei(x−y)ξ dξ = δ(x − y).
Hence
(2π)−d
Z Z
eixξ dξdx = 1.
Proposition 2.1. The Fourier transform of e−itpx is (2π)−d t−d e
iηξ
t
.
Proof.
Z
e−itpx−ipη−ixξ dpdx
=
e
iηξ
t
Z Z
η
ξ
e−it(p+ t )(x+ t ) dpdx.
2
Suppose the variable has a generic name, say x. Then we set
Dx :=
1
∂x .
i
Clearly,
eitDx Φ(x) = Φ(x + t).
2.2
Semiclassical Fourier transformation
If we use the parameter ~, it is natural to use the semiclassical Fourier tranformation
Z
i
F~ f (p) := e− ~ px f (x)dx.
Its inverse is given by
F~−1 g(x) = (2π~)−d
5
Z
i
e ~ xp g(p)dp.
Proposition 2.2. The semiclassical Fourier transformation swaps the position and momentum:
F~−1 x̂F~
=
p̂,
F~−1 p̂F~
=
−x̂.
Proof.
Z
Z
i
i
1
dx
dke ~ px xe− ~ kx Ψ(k)
d
(2π~)
Z
Z
i
i
1
dx dk~i∂k e ~ px e− ~ kx Ψ(k)
=
(2π~)d
Z
Z
i
i
1
dx
dke ~ px e− ~ kx ~i∂k Ψ(k)
=−
(2π~)d
F~−1 x̂F~ Ψ (p) =
F~−1 p̂F~ Ψ
Z
Z
i
i
1
~
(x) =
dx dke ~ px ∂p e− ~ py Ψ(y)
(2π~)d
i
Z
Z
i
i
1
=−
dx
dke ~ px ye− ~ py Ψ(y) =
(2π~)d
(2.3)
(2.4)
=
p̂Ψ(p).
(2.5)
(2.6)
−x̂Ψ(x).
(2.7)
2Hence, for Borel functions
F~−1 f (x̂)F~
= f (p̂),
(2.8)
F~−1 g(p̂)F~
= g(−x̂).
(2.9)
(2.9) can be rewritten as
Z Z
i
1
e ~ (x−y)p g(p)Ψ(y)dydp
d
(2π~)
Z 1
y − x
=
ĝ
Ψ(y)dy.
(2π~)d
~
g(p̂)Ψ)(x) =
2.3
(2.10)
(2.11)
Gaussian integrals
Suppose that ν is positive definite. Then
Z
−1
1
d
1 1
e− 2 x·νx+η·x dx = (2π) 2 (det ν)− 2 e 2 η·ν η .
(2.12)
1
In particular, if f (x) = e− 2 x·νx , its Fourier transform is
−1
d
1
1
fˆ(ξ) = (2π) 2 (det ν)− 2 e− 2 ξ·ν ξ .
6
(2.13)
Consequently,
1
e 2 ∂x ·ν∂x Ψ(x)
=
=
=
1
e− 2 p̂·ν p̂ Ψ(x)
(2.14)
1
d
Z
1
e− 2 (x−y)·ν
(2π)− 2 (det ν)− 2
R − 1 (x−y)·ν −1 (x−y)
e 2
Ψ(y)dy
.
R − 1 y·ν −1 y
dy
e 2
−1
(x−y)
Ψ(y)dy
(2.15)
As a consequence, we obtain the following identity
Z
−1
1
1
(det 2πν)− 2
Ψ(x)e− 2 x·ν x dx
R
−1
1
Ψ(x)e− 2 x·ν x dx
.
=
R − 1 x·ν −1 x
e 2
dx
1
e 2 ∂x ·ν∂x Ψ(0)
=
(2.16)
For example,
e
i~∂x ∂p
−d
Z
i
e ~ (x−y)(p−w) Φ(y, w)dydw.
Φ(x, p) = (2π~)
(2.17)
As an exercise let us check (2.14) for polynomial Ψ. By diagonalizing ν and then changing
the variables, we can reduce ourselves to 1 dimension. Now
1
1
(2π)− 2
Z
1
∞
X
n!
xn−2k ;
2k (n − 2k)!k!
k=0
Z
1
1 2
= (2π)− 2
e− 2 y (y + x)n dy
2
e 2 ∂x xn
=
2
e− 2 (x−y) y n dy
=
∞
X
1
xn−m (2π)− 2
m=0
n!
(n − m)!m!
(2.18)
(2.19)
Z
1
2
e− 2 y y m dy.
Now
1
(2π)− 2
Z
1
(2π)− 2
1
2
e− 2 y y 2k+1 dy
Z
1 2
e− 2 y y 2k dy
=
0,
=
(2k)!
.
2k k!
Indeed, the lhs of (2.20) is
Z ∞
1 k− 12 1 1 2
1
1
1
π − 2 2k
e− 2 y
y2
d y 2 = π − 2 2k Γ k +
,
2
2
2
0
which using
1 (2k)!
1
Γ k+
= π 2 2k
2
2 k!
equals the rhs of (2.20).
Therefore, (2.18)=(2.19).
7
(2.20)
2.4
Gaussian integrals in complex variables
Consider the space Rd ⊕ Rd with the generic variables x, p. It is often natural to identify it
with the complex space Cd by introducing the variables
ai
=
2−1/2 (xi + ipi ),
a∗i
=
2−1/2 (xi − ipi ),
so that
xi = 2−1/2 (a + a∗ ), pi = −i2−1/2 (a − a∗ ).
−d
The Lebesgue measure dxdp will be denoted i
da∗j ∧ daj =
(2.21)
∗
da da. To justify this notation note that
1
dx − idp ∧ dx + idp = idx ∧ dp.
2
Let β be a Hermitian positive quadratic form. We then have
Z
−1
∗
∗
(2πi)−d e−a ·βa+w1 ·a +w2 ·a da∗ da = (det β)−1 ew1 ·β w2 ,
e∂a∗ ·β∂a Φ(a∗ , a)
=
=
(2.22)
Z
∗
∗
−1
(2πi)−d (det β)−1 e−(a −b )·β (a−b) Φ(b∗ , b)db∗ db
R −(a∗ −b∗ )·β −1 (a−b)
e
Φ(b∗ , b)db∗ db
R
.
∗
∗
−1
e−(a −b )·β (a−b) db∗ db
(2.23)
As a consequence, we obtain the following identity
∂a∗ ·β∂a
e
Φ(0, 0)
−d
Z
(2πi) (det β)
Φ(a∗ , a)e−a·β
R
−1 ∗
Φ(a∗ , a)e−a·β a da∗ da
R
.
e−a·β −1 a∗ da∗ da
=
=
−1
−1 ∗
a
da∗ da
(2.24)
As an exercise let us check (2.23) for polynomial Φ. By diagonalizing β and then changing
the variables, we can reduce ourselves to 1 (complex) dimension. Now
∞
X
n!m!
a∗(n−k) am−k ;
(2.25)
(n − k)!(m − k)!k!
k=0
Z
Z
∗
−1
−(a∗ −b∗ )(a−b) ∗n m ∗
−1
(2πi)
e
b b db db = (2πi)
e−b b (a∗ + b∗ )n (a + b)m db∗ db (2.26)
e∂a∗ ∂a a∗n am
=
∞
X
=
a∗(n−k) a(m−l)
k,l=0
n!m!
k!(n − k)!l!(m − l)!
Z
∗
e−b b b∗k bl db∗ db.
Now
(2πi)−1
Z
∗
e−b b b∗k bl db∗ db = k!δkl .
8
(2.27)
Indeed, if we use the polar coordinates with b∗ b = 12 r2 , the lhs of (2.27) becomes
Z ∞
1 21 (k+l)
1 2
e− 2 r
r2
(2π)−1
ei(n−m)φ rdrdφ
2
Z ∞ 0
1 k 1 1 2
= δkl
e− 2 r
r2 d r2
2
2
0
which equals the rhs of (2.27).
Therefore, (2.25)=(2.26).
2.5
Integral kernel of an operator
Every linear operator A on Cn can be represented by a matrix [Aji ].
One would like to generalize this concept to infinite dimensional spaces (say, Hilbert
spaces) and continuous variables instead of a discrete variables i, j. Suppose that a given
vector space is represented, say, as L2 (X), where X is a certain space with a measure.
One often uses the representation of an operator A in terms of its integral kernel X × X 3
(x, y) 7→ A(x, y), so that
Z
AΨ(x) =
A(x, y)Ψ(y)dy.
Note that strictly speaking A(·, ·) does not have to be a function. E.g. in the case X = Rd it
could be a distribution, hence one often says the distributional kernel instead of the integral
kernel. Sometimes A(·, ·) is ill-defined anyway. Below we will describe some situations where
there is a good mathematical theory of integral/distributional kernels.
At least formally, we have
Z
AB(x, y) = A(x, z)B(z, y)dz,
A∗ (x, y) = A(y, x).
2.6
Tempered distributions
The space of Schwartz functions on Rn is defined as
R
S(Rn ) := Ψ ∈ C ∞ (Rn ) : |xα ∇βx Ψ(x)|2 dx < ∞,
Remark 2.3. The definition (2.28) is equivalent to
S(Rn ) = Ψ ∈ C ∞ (Rn ) : |xα ∇βx Ψ(x)| ≤ cα,β ,
α, β ∈ Nn .
α, β ∈ Nn .
(2.28)
(2.29)
S 0 (Rn ) denotes the space of continuous functionals on S(Rn ), ie. S(Rn ) 3 Ψ 7→ hT |Ψi ∈
C belongs to S 0 iff there exists N such that
X Z
21
|hT |Ψi| ≤
|xα ∇βx Ψ(x)|2 dx .
|α|+|β|<N
9
We will use the integral notation
Z
hT |Ψi =
T (x)Ψ(x)dx.
The Fourier transformation is a continuous map from S 0 into itself. We have continuous
inclusions
S(Rn ) ⊂ L2 (Rn ) ⊂ S 0 (Rn ).
Here are some examples of elements of S 0 (R):
Z
δ(t)Φ(t)dt := Φ(0),
Z
Z
(t ± i0)λ Φ(t)dt := lim (t ± i)λ Φ(t)dt.
&0
Theorem 2.4 (The Schwartz kernel theorem). B is a continuous linear transformation
from S(Rd ) to S 0 (Rd ) iff there exists a distribution B(·, ·) ∈ S 0 (Rd ⊕ Rd ) such that
Z
(Ψ|BΦ) = Ψ(x)B(x, y)Φ(y)dxdy, Ψ, Φ ∈ S(Rd ).
Note that ⇐ is obvious. The distribution B(·, ·) ∈ S 0 (Rd ⊕Rd ) is called the distributional
kernel of the transformation B. All bounded operators on L2 (Rd ) satisfy the Schwartz kernel
theorem.
Examples:
(1) e−ixy is the kernel of the Fourier transformation
(2) δ(x − y) is the kernel of identity.
(3) ∂x δ(x − y) is the kernel of ∂x . .
2.7
Hilbert-Schmidt operators
We say that an operator B is Hilbert-Schmidt if
X
X
∞ > TrB ∗ B =
(ei |B ∗ Bei ) =
(Bei |Bei ),
i∈I
i∈I
where {ei }i∈I is an arbitrary basis and the RHS does not depend on the choice of the basis.
Hilbert-Schmidt are bounded.
Proposition 2.5. Suppose that H = L2 (X) for some measure space X. The following
conditions are equivalent
(1) B is Hilbert-Schmidt.
(2) The distributional kernel of B is L2 (X × X).
Moreover, if B, C are Hilbert-Schmidt, then
Z
TrB ∗ C = B(x, y)C(x, y)dxdy.
10
2.8
Trace class operators
B is trace class if
X √
√
∞ > Tr B ∗ B =
(ei | B ∗ Bei ).
i∈I
If B is trace class, then we can define its trace:
X
TrB :=
(ei |Bei ).
i∈I
where again {ei }i∈I is an arbitrary basis and the RHS does not depend on the choice of the
basis.
Trace class operators are Hilbert-Schmidt:
B ∗ B = (B ∗ B)1/4 (B ∗ B)1/2 (B ∗ B)1/4 ≤ (B ∗ B)1/4 kBk(B ∗ B)1/4 = kBk(B ∗ B)1/2 .
Hence
√
TrB ∗ B ≤ kBkTr B ∗ B.
Consider a trace class operator C and a bounded operator B. On the formal level we
have the formula
Z
TrBC = B(y, x)C(x, y)dxdy.
(2.30)
In particular by setting B = 1l, we obtain formally
Z
TrC = C(x, x)dx.
3
3.1
x, p- and Weyl-Wigner quantizations
x, p-quantization
Suppose we look for a linear transformation that to a function b on phase space associates
an operator Op• (b) such that
Op• (f (x)) = f (x̂), Op• (g(p)) = g(p̂).
The so-called x, p-quantization, often used in the PDE community, is determined by the
additional condition
Opx,p (f (x)g(p)) = f (x̂)g(p̂).
Note that
f (x̂)g(p̂) Ψ(x) = (2π~)−d
Z
Z
dp
i(x−y)p
~
Ψ(y).
(3.31)
Hence we can generalize (3.31) for a general function on the phase space b
Z
Z
i(x−y)p
Opx,p (b)Ψ (x) = (2π~)−d dp dyb(x, p)e ~ Ψ(y).
(3.32)
11
dyf (x)g(p)e
In the PDE-community one writes
Opx,p (b) = b(x, ~D).
We also have the closely related p, x-quantization, which satisfies
Opp,x (f (x)g(p)) = g(p̂)f (x̂).
It is given by the formula
p,x
Op
(b)Ψ (x) = (2π~)−d
Z
Z
dp
dyb(y, p)e
i(x−y)p
~
Ψ(y).
Thus the kernel of the operator as x, p- and p, x-quantization is given by:
Z
i(x−y)p
Opx,p (bx,p ) = B, B(x, y) = (2π~)−d dpbx,p (x, p)e ~ ,
Z
i(x−y)p
Opp,x (bp,x ) = B, B(x, y) = (2π~)−d dpbp,x (y, p)e ~ .
Proposition 3.1. We can compute the symbol from the kernel: If (3.34), then
Z
izp
bx,p (x, p) = B(x, x − z)e− ~ dz.
Proof.
B(x, x − z) = (2π~)
−d
Z
e
izp
~
(3.33)
(3.34)
(3.35)
(3.36)
bx,p (x, p)dp.
We apply the Fourier transform. 2
∗
Proposition 3.2. Opx,p (b) = Opp,x (b).
Proposition 3.3. We can go from x, p- to p, x-quantization: If (3.34) and (3.35) hold, then
e−i~Dx Dp bx,p (x, p) = bp,x (x, p).
Proof.
Z
bx,p (x, p)
i
=
B(x, x − z)e− ~ zp dz
Z Z
i
(2π~)−d
bp,x (x − z, w)e ~ z(w−p) dzdw
Z Z
i
(2π~)−d
bp,x (y, w)e ~ (x−y)(w−p) dydw
=
e−i~Dx Dp bp,x (x, p).
=
=
2
Therefore, formally,
Opx,p (b) = Opp,x (b) + O(~).
12
(3.37)
Proposition 3.4. We have the following formula for the symbol of the product: If
Opx,p (b)Opx,p (c) = Opx,p (d),
then
(3.38)
d(x, p) = e−i~Dp1 Dx2 b(x1 , p1 )c(x2 , p2 ) x := x = x , .
1
2
p := p1 = p2 .
Proof.
Opx,p (b)Opx,p (c)(x, y)
Z Z Z
i
b(x, p1 )c(x2 , p2 )e ~ (x−x2 )p1 +(x2 −y)p2 dp1 dx2 dp2
= (2π~)−2d
Z
i
−d
= (2π~)
dp2 e ~ (x−y)p2
Z Z
i
×(2π~)−d
b(x, p1 )c(x2 , p2 )e ~ (x2 −x)(p2 −p1 ) dp1 dx2
Z
i
−d
= (2π~)
dp2 e ~ (x−y)p2 e−i~Dp1 Dx2 b(x1 , p1 )c(x2 , p2 ) x := x = x , .
1
2
p := p1 = p2 .
2
Besides d given by (3.38), let
Opx,p (c)Opx,p (b) = Opx,p (d1 ),
(3.39)
Note that formally
d(x, p)
= b(x, p)c(x, p) − i~∂p b(x, p)∂x c(x, p) + O(~2 ),
d1 (x, p)
= b(x, p)c(x, p) − i~∂x b(x, p)∂p c(x, p) + O(~2 ).
Therefore,
Opx,p (b)Opx,p (c)
=
Opx,p (bc) + O(~),
[Opx,p (b), Opx,p (c)]
=
i~Opx,p ({b, c}) + O(~2 ).
Obviously,
3.2
[x̂, Opx,p (b)]
=
i~Opx,p (∂p b) = i~Opx,p ({x, b}),
[p̂, Opx,p (b)]
= −i~Opx,p (∂x b) = i~Opx,p ({p, b}).
Weyl-Wigner quantization
The definition of the Weyl-Wigner quantization looks like a compromise between the x, p
and p, x-quantizations:
Z
Z
x + y i(x−y)p
Op(b)Ψ (x) = (2π~)−d dp dyb
, p e ~ Ψ(y).
(3.40)
2
13
In the PDE-community it is usually called the Weyl quantization and denoted by
Op(b) = bw (x, ~D).
If Op(b) = B, the kernel of B is given by:
Z
x + y i(x−y)p
−d
,p e ~ .
B(x, y) = (2π~)
dpb
2
Proposition 3.5. We can compute the symbol from the kernel:
Z z
z − izp
b(x, p) = B x + , x −
e ~ dz.
2
2
Proof.
z
z
B(x + , x − ) = (2π~)−d
2
2
Z
e
izp
~
(3.41)
b(x, p)dp.
We apply the Fourier transform. 2
Usually, b is called in the PDE community the Weyl symbol and in the quantum physics
community the Wigner function.
1 2
d
Let P0 be the orthogonal projection onto the normalized vector π − 4 e− 2 x . The integral
kernel of P0 equals
d
1 2
1 2
P0 (x, y) = π − 2 e− 2 x − 2 y .
Its various symbols equal
x, p-symbol:
p, x-symbol:
Weyl-Wigner symbol:
− 21 p2 −ix·p
,
d
2
− 12 x2 − 21 p2 +ix·p
,
d
2
− 12 x2 − 21 p2
d
1
2
2 2 e− 2 x
2 e
2 e
.
Proposition 3.6. We can go from the x, p- to the Weyl quantization:
Opx,p (bx,p ) = Op(b), then
if
i
e 2 ~Dx Dp b(x, p) = bx,p (x, p).
(3.42)
Consequently,
bx,p = b + O(~).
3.3
Star product
Proposition 3.7. We have the following formula for the symbol of the product:
if
Op(b)Op(c) = Op(d), then
i
d(x, p) = e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) b(x1 , p1 )c(x2 , p2 ) x := x = x , .
1
2
p := p1 = p2 .
14
(3.43)
Consequently,
1
Op(b)Op(c) + Op(c)Op(b) = Op(bc) + O(~2 ),
2
[Op(b), Op(c)] = i~Op({b, c}) + O(~3 ),
if suppb ∩ suppc = ∅,
then
Op(b)Op(c)
O(~∞ ).
=
(3.44)
(3.45)
(3.46)
Often one denotes d in (3.43) by b ∗ c and calls the star product or the Moyal product of
b and c.
Proposition 3.8. If h is a polynomial of degree ≤ 1, then
1
Op(h)Op(b) + Op(b)Op(h)
2
=
Op(bh),
Proof. Consider for instance h = x.
i
e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) x1 b(x2 , p2 ) x := x = x ,
1
2
p := p1 = p2 .
i
e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) b(x1 , p1 )x2 x := x = x ,
1
2
= xb(x, p) +
i~
∂p b(x, p),
2
= xb(x, p) −
i~
∂p b(x, p).
2
p := p1 = p2 .
2
Consequently,
2
p̂ − A(x̂)
Op aij xi xj + 2bij xi pj + cij pi pj
=
2 Op p − A(x)
,
=
aij x̂i x̂j + bij x̂i p̂j + bij p̂j x̂i + cij p̂i p̂j .
Proposition 3.9. Let h be a polynomial of degree ≤ 2. Then
(1)
[Op(h), Op(b)] = i~Op({h, b}).
(3.47)
(2) Let x(t), p(t) solve the Hamilton equations with the Hamiltonian h. Then the affine
symplectic transformation
rt x(0), p(0) = x(t), p(t)
satisfies
it
it
e ~ Op(h) Op(b)e− ~ Op(h) = Op(b ◦ rt ).
15
Proof.
i
e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) h(x1 , p1 )b(x2 , p2 ) x := x = x ,
1
2
p := p1 = p2 .
i~
= h(x, p)b(x, p) + (Dp h(x, p)Dx b(x, p) − Dp h(x, p)Dx b(x, p) ,
2
(i~)2
(Dp1 Dx2 − Dx1 Dp2 )2 h(x1 , p1 )b(x2 , p2 ) x := x = x ,
+
1
2
8
p := p1 = p2 .
When we swap h and b, we obtain the same three terms except that the second has the
opposite sign. This proves (1).
To prove (2) note that
d
b ◦ rt = {h, b ◦ rt }
dt
[Op(h), Op(b ◦ rt )] = i~Op({h, b ◦ rt }).
it
it
Now, e− ~ Op(h) Op(b ◦ rt )e ~ Op(h) = Op(b) and
t=0
=
it
d −it Op(h)
Op(b ◦ rt )e ~ Op(h)
e ~
dt
i
it
−it
d
e ~ Op(h) − [Op(h), Op(b ◦ rt )] + Op
b ◦ rt e ~ Op(h) = 0
~
dt
2
3.4
Weyl operators
Proposition 3.10 (Baker-Campbell-Hausdorff formula). Suppose that
[A, B], A = [A, B], B = 0.
Then
1
eA+B = eA eB e− 2 [A,B] .
Proof. We will show that for any t ∈ R
1 2
et(A+B) = etA etB e− 2 t
[A,B]
.
First, using the Lie formula, we obtain
etA Be−tA
=
∞ n
X
t
adnA (B)
n!
n=0
=
B + t[A, B].
16
(3.48)
Now
d tA tB − 1 t2 [A,B]
e e e 2
dt
1 2
AetA etB e− 2 t
=
[A,B]
1 2
+etA BetB e− 2 t
[A,B]
1 2
−etA etB t[A, B]e− 2 t
1 2
(A + B)etA etB e− 2 t
=
[A,B]
[A,B]
.
Besides, (3.48) is true for t = 0. 2
Let ξ = (ξ1 , . . . , ξd ), η = (η1 , . . . , ηd ) ∈ Rd . Clearly,
[ξi x̂i , ηj p̂j ] = i~ξi ηi .
Therefore,
i~
e− 2 ξi ηi ei(ξi x̂i +ηi p̂i )
= e−i~ξi ηi eiηi p̂i eiξi x̂i .
eiξi x̂i eiηi p̂i
=
The operators ei(ξi x̂i +ηi p̂i ) are sometimes called Weyl operators. They satisfy the relations
that involve the symplectic form:
0
0
0
0
0
0
i~
ei(ξi x̂i +ηi p̂i ) ei(ξi x̂i +ηi p̂i ) = e− 2 (ξi ηi −ηi ξi ) ei (ξi +ξi )x̂i +(ηi +ηi )p̂i
(3.49)
They translate the position and momentum:
i
i
i
~ (−y p̂+wx̂)
i
~ (y p̂−wx̂)
e ~ (−yp̂+wx̂) x̂e ~ (yp̂−wx̂)
e
3.5
p̂e
=
x̂ − y,
=
p̂ − w.
Weyl-Wigner quantization in terms of Weyl operators
Note that
i
i
ei(ξi x̂i +ηi p̂i ) = e 2 ξi x̂i eiηi p̂i e 2 ξi x̂i .
Hence the integral kernel of e
i(ξi x̂i +ηi p̂i )
(2π~)−d
Z
is
1
1
i
dpei( 2 ξi xi +ηi pi + 2 ξi yi )+ ~ (xi −yi )pi .
Therefore,
Op(ei(ξi xi +ηi pi ) ) = ei(ξi x̂i +ηi p̂i ) .
More generally, if f is a function on R, then
=
(2π)−1
Z
f (ξi x̂i + ηi p̂i )
=
(2π)−1
Z
f (ξi xi + ηi pi )
17
Ff (t)ei(ξi x̂i +ηi p̂i )t dt,
Ff (t)ei(ξi xi +ηi pi )t dt.
(3.50)
Therefore,
Op f (ξi xi + ηi pi ) = f (ξi x̂i + ηi p̂i ).
Every function b on Rd ⊕ Rd can be written in terms of its Fourier transform:
Z Z Z Z
b(x, p) = (2π)−2d
ei(xi −yi )ξi +i(pi −wi )ηi b(y, w)dydwdξdη.
(3.51)
(3.52)
This leads to
Op(b) = (2π)−2d
Z Z Z Z
ei(x̂i −yi )ξi +i(p̂i −wi )ηi b(y, w)dydwdξdη,
(3.53)
which can be treated as an alternative definition of the Weyl-Wigner quantization.
3.6
Classical and quantum mechanics over a symplectic vector space
A symplectic form is a nondegenerate antisymmetric form. A vector space equipped with a
symplectic form is called a symplectic vector space. It has to have an even dimension.
Let R2d be an even dimensional vector space. Let φj , j = 1, . . . , 2d denote the canonical
coordinates in R2d . Let ω = [ωij ] be a symplectic form. We will denote by [ω ij ] the inverse
of [ωij ].
With every symplectic space we have a classical system. Indeed, we can define the
Poisson bracket for functions b, c on R2d :
{b, c} = ∂φi bω ij ∂φj c.
Proposition 3.11. In every symplectic space we can choose a basis φi = xi , φd+i = pi so
that the Poisson bracket has the form (1.1), that is
ωi,i+d = 1,
ωi+d,i = −1.
We also have a quantum system with operators φ̂j , j = 1, . . . , 2d satisfying
[φ̂j , φ̂k ] = iω jk 1l.
(3.54)
We say that a linear transformation r = [rij ] is symplectic if it preserves the symplectic
form. Explicitly, r# ωr = ω, or
rip ωpq rjq = ωij
The set of such transformations is denoted Sp(R2d ). It is a Lie group.
We say that a linear transformation b = [bji ] is infinitesimally symplectic if it infinitesimally preserves the symplectic form. In other words, 1l + b is for small approximately
symplectic. Explicitly, b# ω + ωb = 0, or
bpi ωpj + ωip bpj = 0.
The set of such transformations is denoted sp(R2d ). It is a Lie algebra.
18
Proposition 3.12. b is an infinitesimally symplectic transformaton iff c = ω −1 b is symmetric, so that bij = ω ik ckj . Then
H=
1
1
cjk φj φk , Ĥ = Op(H) = cjk φ̂j φ̂k
2
2
is the corresponding classical and quantum Hamiltonian. Let [rij (t)] be the corresponding dynamics, which is a 1-parameter group in Sp(R2d ). Then the classical and quantum dynamics
generated by this Hamiltonian are given by the flow r(t):
φj (t) = rkj (t)φk (0),
3.7
φ̂j (t) = rkj (t)φ̂k (0),
Weyl quantization for a symplectic vector space
Let φ̂i satisfy the relations (3.54). For ζ = (ζ1 , . . . , ζ2d ) ∈ R2d set
X
φ̂·ζ :=
φ̂i ζi , W (ζ) := eiζ·φ̂ .
i
Then
i
W (ζ)W (θ) = e− 2 ζ·ωθ W (ζ + θ).
For a function b on R2d we can define its Weyl-Wigner quantization:
Z Z
i
−2d
Op(b) := (2π)
ei(φ̂i −ψi )·ζ b(ψ)dψdζ.
Note that
Op eiφ·ζ = eiφ̂·ζ .
(3.55)
More generally, for any Borel function f
Op f (φ·ζ) = f (φ̂·ζ).
(3.56)
1 X
φ̂ζσ(i) · · · φ̂ζσ(n) .
Op φζ1 · · · φζn =
n!
(3.57)
Op (φζ)n = (φ̂ζ)n .
(3.58)
Proposition 3.13.
σ∈Sn
Proof. We have
This is a special case of (3.57), it is also seen directly from (3.55) by expanding into a power
series. Let t1 , . . . , tn ∈ R. By (3.58),
Op (t1 φ·ζ1 + · · · + tn φ·ζn )n = (t1 φ̂·ζ1 + · · · + tn φ̂·ζn )n .
(3.59)
The coefficient at t1 · · · tn on both sides is
Op(n!φ·ζ1 · · · φ·ζn ) =
X
σ∈Sn
2
19
φ̂·ζσ−1 (1) · · · φ̂·ζσ−1 (n) .
3.8
Positivity
Clearly,
Op(b)∗ = Op(b).
Therefore, b is real iff Op(b) is Hermitian. What about positivity? We will see that there is
no implication in either direction between the positivity of b and of Op(b).
We have
(x̂ − ip̂)(x̂ + ip̂) = x̂2 + p̂2 − ~ ≥ 0.
Therefore
Op(x2 + p2 − ~) ≥ 0,
2
(3.60)
2
even though x + p − ~ is not everywhere positive.
The converse is more complicated. Consider the generator of dilations
A :=
i
1
(x̂p̂ + p̂x̂) = x̂p̂ − = Op(xp).
2
2
Its name comes from the 1-parameter group it generates:
eitA Φ(x) = et/2 Φ(et x).
Note that sp A = R. Indeed, A preserves the direct decomposition L2 (R) = L2 (0, ∞) ⊕
L2 (−∞, 0). We will show that the spectrum of A restricted to each of these subspaces
is R. Consider the unitary operator U : L2 (0, ∞) → L2 (R) given by U Φ(s) = es/2 Φ(es )
with the inverse U ∗ Ψ(x) = x−1/2 Ψ(log x). Then U ∗ p̂U = A. But sp p̂ = R. Therefore,
sp A2 = (sp A)2 = [0, ∞[.
We have A2 = Op(b), where
i~
b(x, p) = e 2 (Dp1 Dx2 −Dx1 Dp2 ) x1 p1 x2 p2 x := x = x ,
1
2
p := p1 = p2 .
2
~
2Dp1 Dx2 Dx1 Dp2 x1 p1 x2 p2 x := x = x ,
= x2 p 2 +
1
2
4·2
p := p1 = p2 ,
= x2 p 2 +
~2
.
4
Hence
Op(x2 p2 ) = A2 −
~2
.
4
Therefore Op(x2 p2 ) is not a positive operator even though its symbol is positive
3.9
Parity operator
Define the parity operator
IΨ(x) = Ψ(−x).
20
(3.61)
More generally, set
i
i
I(y,w) := e ~ (−yp̂+wx̂) Ie ~ (yp̂−wx̂) .
(3.62)
Clearly,
2i
I(y,w) Ψ(x) = e ~ w·(x−y) Ψ(2y − x).
Let δ(y,w) denote the delta function at (y, w) ∈ Rd ⊕ Rd .
Proposition 3.14.
Op (π~)d δ(0,0) = I.
(3.63)
Op (π~)d δ(y,w) = I(y,w) .
(3.64)
More generally,
Proof.
d
Op (π~) δ(0,0) (x, y)
−d
Z
=
2
=
2−d δ
δ
x + y
2
x + y
2
i
, ξ e ~ (x−y)·ξ dξ
= δ(x + y).
To see the last step we substitute y2 = ỹ below and evaluate the delta function:
Z Z x + y
x
δ
Φ(y)dy = δ
+ ỹ Φ(2ỹ)2d dỹ = 2d Φ(−x).
2
2
(3.65)
2
Theorem 3.15. Let Op(b) = B.
(1) If b ∈ L1 (Rd ⊕ Rd ), then B is a compact operator. In terms of an absolutely norm
convergent integral, we can write
Z
B = (π~)−d I(x,p) b(x, p)dxdp.
(3.66)
Hence,
kBk ≤ (π~)−d kbk1 .
(3.67)
(2) If B is trace class, then b is continuous, vanishes at infinity and
b(x, p) = 2d TrI(x,p) B.
Hence
|b(x, p)| ≤ 2d Tr|B|.
Proof. Obviously,
Z
b=
b(x, p)δx,p dxdp.
21
(3.68)
Hence
Z
Op(b)
=
=
b(x, p)Op(δx,p )dxdp
Z
−d
(π~)
b(x, p)I(x,p) dxdp.
Next,
Z
b(x, p)
=
δ(x,p) (y, w)b(y, w)dydw
=
(2π~)d TrOp(δ(x,p) )Op(b)
=
2d TrI(x,p) Op(b).
2
3.10
Special classes of symbols
Proposition 3.16. The following conditions are equivalent
(1) B is continuous from S to S 0 .
(2) The x, p-symbol of B is Schwartz.
(3) The p, x-symbol of B is Schwartz.
(4) The Weyl-Wigner symbol of B is Schwartz.
Proof. By the Schwartz kernel theorem (1) is equivalent to B having the kernel in S 0 . The
formulas () involve only partial Fourier transforms and some constant coefficients. 2
Proposition 3.17. The following conditions are equivalent
(1) B is Hilbert-Schmidt.
(2) The x, p-symbol of B is L2 .
(3) The p, x-symbol of B is L2 .
(4) The Weyl-Wigner symbol of B is L2 .
Moreover, if b, c are L2 , then
TrOpx,p (b)∗ Opx,p (c) = TrOp(b)∗ Op(c) = (2π~)−d
Z
b(x, p)c(x, p)dxdp.
(3.69)
One often uses (3.69) when B = Op(b) is, say, bounded and describes an observable, and
C = Op(c) is trace class, and describes a density matrix. Setting b(x, p) = 1 we formally
obtain
Z
x,p
−d
TrOp(c) = TrOp (c) = (2π~)
c(x, p)dxdp.
22
With the x, p-quantization we can compute the so-called marginals involving the position
and momentum:
Z
Trf (x̂)Opx,p (c) = (2π~)−d f (x)c(x, p)dxdp,
Z
Trg(p̂)Opx,p (c) = (2π~)−d g(p)c(x, p)dxdp.
With the Weyl-Wigner quantization we have much more possibilities: eg. for any α
Z
Trf (cos αx̂ + sin αp̂)Op(c) = (2π~)−d f (cos αx + sin αp)c(x, p)dxdp.
4
4.1
Coherent states
General coherent states
Fix a normalized vector Ψ ∈ L2 (Rd ). The family of coherent vectors associated with the
vector Ψ is defined by
i
Ψ(y,w) := e ~ (−yp̂+wx̂) Ψ,
(y, w) ∈ Rd ⊕ Rd .
The orthogonal projection onto Ψ(y,w) , called the coherent state, will be denoted
i
i
P(y,w) := |Ψ(y,w) )(Ψ(y,w) | = e ~ (−yp̂+wx̂) |Ψ)(Ψ|e ~ (yp̂−wx̂) .
It is natural to assume that
Ψ|x̂Ψ = 0,
Ψ|p̂Ψ = 0.
This assumption implies that
Ψ(y,w) |x̂Ψ(y,w) = y,
Ψ(y,w) |p̂Ψ(y,w) = w.
Note however that we will not use the above assumption in this section.
Explicitly,
i
=
e ~ (w·x− 2 y·w) Ψ(x − y),
P(y,w) (x1 , x2 )
=
Ψ(x1 − y)Ψ(x2 − y)e ~ (x1 −x2 )·w .
i
Theorem 4.1.
−d
(2π~)
Proof.
1
Ψ(y,w) (x)
Z
P(y,w) dydw = 1l.
Let Φ ∈ L2 (Rd ). Then
Z Z
(Φ|P(y,w) Φ)dydw
Z Z Z Z
i
=
Φ(x1 )Ψ(x1 − y)Ψ(x2 − y)e ~ (x1 −x2 )·w Φ(x2 )dx1 dx2 dydw
Z Z
= (2π~)d
Φ(x)Ψ(x − y)Ψ(x − y)Φ(x)dxdy = (2π~)d kΦk2 kΨk2 .
23
(4.70)
4.2
Contravariant quantization
Let b be a function on te phase space. We define its contravariant quantization by
Z
Opct (b) := (2π~)−d P(x,p) b(x, p)dxdp.
(4.71)
If B = Opct (b), then b is called the contravariant symbol of B.
We have
R
(1) |TrOpct (b)| ≤ (2π~)−d |b(x, p)|dxdp;
(2) kOpct (b)k ≤ supx,p |b(x, p)|;
(3) Opct (1) = 1l;
(4) Opct (b)∗ = Opct (b).
(5) Let b ≥ 0. Then Opct (b) ≥ 0.
4.3
Covariant quantization
The covariant quantization is the operation dual to the contravariant quantization. Strictly
speaking, the operation that has a natural definition and good properties is not the covariant
quantization but the covariant symbol of an operator.
Let B ∈ B(H). Then we define its covariant symbol by
b(x, p) := TrP(x,p) B = Ψ(x,p) |BΨ(x,p) .
B is then called the covariant quantization of b and is denoted by
Opcv (b) = B.
(1) Opcv (1) = 1l,
(2) Opcv (b)∗ = Opcv (b).
(3) kOpcv (b)k ≥ supx,p |b(x, p)|.
(4) Let Opcv (b) ≥ 0. Then bcv ≥ 0.
R
(5) TrOpcv (b) = (2π~)−d b(x, p)dxdp.
4.4
Connections between various quantizations
Let us compute various symbols of P(y,w) :
covariant symbol(x, p)
= |(Ψ|Ψ(y−x,w−p) |2 ,
Weyl symbol(x, p)
=
2d (Ψ(y−x,w−p) |IΨ(y−x,w−p) ),
contravariant symbol(x, p)
=
(2π~)d δ(x − y)δ(p − w).
Let us now show how to pass between the covariant, Weyl-Wigner and contravariant
quantization. Note that there is a preferred direction: from contravariant to Weyl, and then
from Weyl-Wigner to covariant. Going back is less natural.
24
Let
Opct (bct ) = Op(b) = Opcv (bcv ).
Then
−d
b(x, p)
=
(π~)
bcv (x, p)
=
(π~)−d
bcv (x, p)
=
Z
bct (y, w)(Ψ(y−x,w−p) |IΨ(y−x,w−p) )dydw,
Z
(2π~)−d
b(y, w)(Ψ(−y+x,−w+p) |IΨ(−y+x,−w+p) )dydw,
Z
2
bct (y, w)(Ψ|Ψ(y−x,w−p) ) dydw.
We have
TrOpcv (a)Opct (b) = (2π~)−d
Z
a(x, p)b(x, p)dxdp.
(4.72)
Indeed, let A = Opcv (a). Then the lhs of (4.72) is
Z
TrA(2π~)−d b(x, p)|Ψ(x,p) )(Ψ(x,p) |dxdp
Z
−d
= (2π~)
(Ψ(x,p) |A|Ψ(x,p) )b(x, p)dxdp,
which is the rhs of (4.72).
4.5
Gaussian coherent vectors
Consider the normalized gaussian vector scaled appropriately with the Planck constant
1
d
2
Ω(x) = (π~)− 4 e− 2~ x .
(4.73)
The corresponding coherent vectors are equal to
d
i
i
1
2
Ω(y,w) (x) = (π~)− 4 e ~ w·x− 2~ y·w− 2~ (x−y) .
(4.74)
In the literature, when one speaks about coherent states, one has usually in mind (4.74).
They are also called Gaussian or Glauber’s coherent states.
Z Z
2
2
2
2
1
~
1
b(x, p) =
bct (y, w)(π~)−d e− ~ (x−y) − ~ (p−w) dydw, b = e− 4 (Dx +Dp ) bct ;
Z Z
2
2
2
2
1
1
~
cv
b (x, p) =
b(y, w)(π~)−d e− ~ (x−y) − ~ (p−w) dydw, bcv = e− 4 (Dx +Dp ) b;
Z Z
2
2
2
2
1
1
~
bcv (x, p) =
bct (y, w)(2π~)−d e− 2~ (x−y) − 2~ (p−w) dydw, bcv = e− 2 (Dx +Dp ) bct .
In the case of Gaussian states, there are several alternative names of the covariant and
contravariant symbol of an operator:
(1) For contravariant symbol:
25
(i) anti-Wick symbol,
(ii) Glauber-Sudarshan function,
(iii) P-function;
(2) For covariant symbol:
(i) Wick symbol,
(ii) Husimi or Husimi-Kano function,
(iii) Q-function.
We will use the terms Wick/anti-Wick quantization/symbol.
One can distinguish 5 most natural quantizations. Their respective relations are nicely
described by the following diagram, called sometimes the Berezin diagram:
anti-Wick
quantization
 ~ 2 2
 − 4 (Dx +Dp )
ye
p, x
quantization
i~
e 2 Dx ·Dp
−→
Weyl-Wigner
quantization
 ~ 2 2
 − 4 (Dx +Dp )
ye
i~
e 2 Dx ·Dp
−→
x, p
quantization
Wick
quantization
4.6
Creation and annihilation operator
Set
ai
=
(2~)−1/2 (xi + ipi ),
a∗i
=
(2~)−1/2 (xi − ipi ).
We have
i
{ai , a∗j } = − δij .
~
~1/2
~1/2
xi = 1/2 (ai + a∗i ), pi = 1/2 (ai − a∗i ).
(4.75)
2
i2
In this way, the classical phase space Rd ⊕ Rd has been identified with the complex space
Cd . The Lebesgue measure has also a complex notation:
~d ∗
da da = dxdp.
id
To justify the notation (4.76) we write in terms of differential forms:
da∗j ∧ daj =
1
dx − idp ∧ dx + idp = i~−1 dx ∧ dp.
2~
26
(4.76)
On the quantum side we introduce the operators
âi
=
(2~)−1/2 (x̂i + ip̂i ),
â∗i
=
(2~)−1/2 (x̂i − ip̂i ).
We have
[âi , â∗j ] = δij .
~1/2
~1/2
∗
(â
+
â
),
p̂
=
(âi − â∗i ).
i
i
i
21/2
i21/2
Let b, b∗ be classical variables obtained from y, w, as above. Note that
(4.77)
x̂i =
i
e ~ (−yp̂+wx̂) = e(−b
∗
â+bâ∗ )
.
(4.78)
We have
i
i
i
e ~ (−yp̂+wx̂) x̂ = (x̂ + y)e ~ (−yp̂+wx̂) ,
e(−b
∗
∗
∗
â+bâ ) ∗
â = (â∗ + b∗ )e(−b
∗
â+bâ )
i
e ~ (−yp̂+wx̂) p̂ = (p̂ + w)e ~ (−yp̂+wx̂) ,
∗
e(−b
,
∗
â+bâ )
∗
â = (â + b)e(−b
∗
â+bâ )
.
(4.79)
(4.80)
Recall that in the real notation we had coherent vectors
i
Ωy,w := e ~ (−yp̂+wx̂) Ω.
(4.81)
In the complex notation they become
Ωb := e(−b
x2
∗
â+bâ∗ )
Ω.
(4.82)
1
Using Ω(x) = e− 2~ and âi = (2~)− 2 (x̂i + ~∂xi ) we obtain
âi Ω = 0.
This justifies the name “annihilation operators” for âi . More generally, by (4.80),
âj Ωb = bj Ωb ,
Note that the identity (4.70) can be rewritten as
Z
−d
1l = (2πi)
|Ωa )(Ωa |da∗ da.
4.7
Quantization by an ordering prescription
Consider a polynomial function on the phase space:
X
w(x, p) =
wα,β xα pβ .
α,β
27
(4.83)
It is easy to describe the x, p and p, x quantizations of w in terms of ordering the positions
and momenta:
X
Opx,p (w) =
wα,β x̂α p̂β ,
α,β
Opp,x (w)
X
=
wα,β p̂β x̂α .
α,β
The Weyl quantization amounts to the full symmetrization of x̂i and p̂j , as described in
(3.57).
We can also rewrite the polynomial (4.83) in terms of ai , a∗i by inserting (4.75). Thus
we obtain
X
w(x, p) =
w̃γ,δ a∗γ aδ =: w̃(a∗ , a).
(4.84)
γ,δ
Then we can introduce the Wick quantization
X
∗
Opa ,a (w) =
w̃γ,δ â∗γ âδ
(4.85)
γ,δ
and the anti-Wick quantization
∗
Opa,a (w) =
X
w̃γ,δ âδ â∗γ .
(4.86)
γ,δ
Theorem 4.2. (1) The Wick quantization coincides with the covariant quantization for
Gaussian coherent states.
(2) The anti-Wick quantization coincides with the contravariant quantization for Gaussian
coherent states.
Proof. (1)
∗
(Ω(x,p) |Opa
,a
(w)Ω(x,p) )
Ωa |
=
X
w̃γ,δ â∗γ âδ Ωa )
γ,δ
X
=
w̃γ,δ a∗γ aδ
γ,δ
= w(x, p).
(2)
a,a∗
Op
(w)
=
X
−d
δ
w̃γ,δ â (2πi)
Z
|Ωa )(Ωa |da∗ daâ∗γ
γ,δ
=
(2πi)−d
XZ
w̃γ,δ aδ a∗γ |Ωa )(Ωa |da∗ da
γ,δ
=
(2π~)−d
Z
w(x, p)|Ω(x,p) )(Ω(x,p) |dxdp.
28
2
The Wick quantization is widely used, especially for systems with an infinite number of
degrees of freedom. Note the identity
∗
Ω|Opa
4.8
,a
(w)Ω) = w̃(0, 0).
(4.87)
Connection between the Wick and anti-Wick quantization
As described in equation (4.83), there are two natural ways to write the symbol of the Wick
(or anti-Wick) quantization. We can either write it in terms of x, p, or in terms of a∗ , a. In
the latter notation we decorate the symbol with a tilde.
Let
∗
∗
∗
∗
Opa,a (wa,a ) = Opa ,a (wa ,a ).
Then
∗
wa
∗
w̃a
,a
,a
(x, p)
(a∗ , a)
~
2
∗
2
e 2 (∂x +∂p ) wa,a (x, p)
Z Z
1
∗
− 2~
(x−y)2 +(p−w)2
−d
= (2π~)
e
wa,a (y, w)dydw,
=
(4.88)
∗
e∂a∗ ∂a w̃a,a (a∗ , a)
Z Z
∗
∗
∗
= (2πi)−d
e−(a −b )(a−b) w̃a,a (b∗ , b)db∗ db..
=
(4.89)
(4.88) was proven before. To see that (4.89) and (4.88) are equivalent we note that
∂a
=
∂a∗
=
~1/2
(∂x + i∂p ),
21/2
~1/2
(∂x − i∂p ),
21/2
hence
~ 2
(∂ + ∂p2 ).
2 x
One can also see (4.89) directly. To this end it is enough to consider a∗n am (a and a∗ are
now single variables). To perform Wick ordering we need to make all possible contractions.
Each contraction involves a pair of two elements: one from {1, . . . , n} and the other from
{1, . . . , m}. The number of possible k-fold contractions is
∂a∗ ∂a =
m!
1
n!
m!
n!
k! =
.
k!(n − k)! k!(m − k)!
k! (n − k)! (m − k)!
But
e∂a∗ ∂a a∗n am =
∞
X
1
n!
m!
a∗(n−k)
am−k .
k! (n − k)!
(m − k)!
k=0
29
4.9
Wick symbol of a product
Suppose that
∗
Opa
,a
(w) = Opa
Then
w̃(a∗ , a) = e
∂a2 ∂a∗
1
∗
,a
∗
(w2 )Opa
,a
(w1 ).
w̃2 (a∗2 , a2 )w̃1 (a∗1 , a1 )
a = a2 = a1 .
(4.90)
(Clearly a = a2 = a1 implies a∗ = a∗2 = a∗1 ). This follows essentially by the same argument
as the one used to show (4.89). Using (2.14), one can rewrite (4.90) as an integral:
Z Z
db∗ db
w̃(a∗ , a) =
.
(4.91)
w̃2 (a∗ , b)w̃1 (b∗ , a)
(2πi)d
We also have a version of (4.90) for an arbitrary number of factors. Let
Opa
∗
,a
∗
(w) = Opa
,a
(wn ) · · · Opa
∗
,a
(w1 ).
Then
∂ak ∂a∗j w̃n (a∗n , an ) · · · w̃1 (a∗1 , a1 )
a = an = · · · = a1 .
∗
∗
Ω|Opa ,a (wn ) · · · Opa ,a (w1 )Ω
X
exp
∂ak ∂a∗j w̃n (a∗n , an ) · · · w̃1 (a∗1 , a1 )
0 = an = · · · = a1 .
w̃(a∗ , a) = exp
X
k>j
(4.92)
Consequenly,
=
k>j
We can rewrite (4.92) as an integral:
w̃(a∗ , a)
(4.93)
Z
Z
n−1
n−1
X
Y dbj+1 db∗j
.
(bj+1 − bj )(b∗j − a∗ ) w̃n (b∗n , bn ) · · · w̃1 (b∗1 , b1 )
= · · · exp
(2πi)d
j=1
j=1
Indeed, introduce new variables
cn := an − an−1 ,
cn−1 = an−1 − an−2 , . . . ,
c1 = a1 .
Then
∂cn = ∂an ,
∂cn−1 = ∂an + ∂an−1 , . . . ,
∂c1 = ∂an + · · · + ∂a1 .
Therefore,
X
k>j
∂ak ∂a∗j =
n−1
X
j=1
30
∂cj+1 ∂a∗j .
Hence
exp
X
∂ak ∂a∗j Ψ(a∗n , an , . . . , a∗1 , a1 )
(4.94)
k>j
Z
=
Z
···
exp
n−1
n−1
X
Y dbj+1 db∗j
(bj+1 − bj − aj+1 + aj )(b∗j − a∗j ) Ψ(b∗n , bn , . . . , b∗1 , b1 )
(2πi)d
j=1
j=1
Finally, we insert in (4.94) an = · · · = a1 = a, a∗n = · · · = a∗1 = a∗ .
5
5.1
Pseudodifferential calculus with uniform symbol classes
Gaussian dynamics on uniform symbol classes
We will denote by S(Rn ) the space of b ∈ C ∞ (Rn ) such that
|∂xα b| ≤ Cα .
Note that S(Rn ) has the structure of a Frechet space.
i
Theorem 5.1. Let ν be a quadratic form. Then e 2 DνD is bounded on S(Rn ). Then, for ν
in a bounded set of quadratic forms, there exist C and m such that
i
sup |e 2 DνD b(x)| ≤ C sup |∂β b(x)|.
(5.95)
|β|≤m
i
Consequently, e 2 DνD is bounded on S(Rn ).
5.2
The boundedness of quantized uniform symbols
Let us set ~ = 1.
We will denote by S(Rd ⊕ Rd ) the space of b ∈ C ∞ (Rd ⊕ Rd ) such that
|∂xα ∂pβ b| ≤ Cα,β .
Theorem 5.2 (The Calderon-Vaillancourt Theorem). If b ∈ S(Rd ⊕ Rd ), then Opx,p (b) is
bounded and
X
kOpx,p (b)k ≤ C(n)
sup |∂xα ∂pβ b(x, p)|.
|α|+|β|≤N (n)
There exists also a version of the Calderon-Vaillancourt Theorem with the Weyl quantization replacing the x, p-quantization.
We will write adB (A) := [B, A].
Theorem 5.3. The following conditions are equivalent:
(1)
β
2
d
adα
x̂ adp̂ B ∈ B(L (R )), α, β.
31
(2) B = Opx,p (b), b ∈ S(Rd ⊕ Rd ),
(3) B = Op(b), b ∈ S(Rd ⊕ Rd ).
The implication (1)⇒(2) is called the Beals Criterion. The Calderon-Vaillancourt Theorem implies (1)⇐(2). Let us denote by Ψ the set of operators described in Theorem 5.3.
Theorem 5.4. (1) Ψ is an algebra.
(2) Let B ∈ Ψ be boundedly invertible. Then B −1 ∈ Ψ.
(3) Let f be a function holomorphic on sp B, where B ∈ Ψ. Then f (B) ∈ Ψ.
(4) Let f be a function smooth on sp B, where B = B ∗ ∈ Ψ. Then f (B) ∈ Ψ.
5.3
Semiclassical calculus
We go back to ~. We will write Op~ for the quantization depending on ~.
Theorem 5.5. Let a, b ∈ S(Rd ⊕ Rd ). Then there exist c0 , . . . , cn ∈ S and ~ 7→ r~ ∈
S(Rd ⊕ Rd ) such that
Op~ (a)Op~ (b) =
n
X
~2j Op~ (c2j ) + ~2n+2 Op~ (r~ ),
j=0
α β k ∂x ∂p ∂~ r~ ≤ Cα,β,k .
Besides,
c0 = ab, c1 =
i
{a, b}.
2
If in addition a or b is 0 on an open set Θ ⊂ Rd ⊕ Rd , then so are c0 , . . . , cn .
For a complex function b let ran(b) denote the closure of the image of b.
Theorem 5.6. Let b ∈ S and 0 6∈ ran(b). Then for small enough ~ the operator Op~ (b) is
invertible and there exist c0 , c2 , . . . , c2n ∈ S and ~ 7→ r~ ∈ S such that
Op~ (b)−1 =
n
X
~2j Op~ (c2j ) + ~2n+2 Op~ (r~ ),
j=0
α β k ∂x ∂p ∂~ r~ ≤ Cα,β,k .
Besides,
c0 = b−1 .
As a corollary of the above
theorem, for any neighborhood of ran(b) there exists ~0 such
that, for |~| ≤ ~0 , sp Op~ (b) is contained in this neighborhood.
32
Theorem 5.7. Let b ∈ S and f be a function holomorphic on a neighborhood
of the image
of b. Then for small enough ~ the function f is defined on sp Op~ (b) and there exist
c0 , c2 , . . . , c2n ∈ S and ~ 7→ r~ ∈ S such that
n
X
f Op~ (b) =
~2j Op~ (c2j ) + ~2n+2 Op~ (r~ ),
j=0
α β k ∂x ∂p ∂~ r~ ≤ Cα,β,k .
Besides,
c0 = f ◦ b.
If b is real and smooth, then the same conclusion holds.
Theorem 5.8. Let c ∈ S(Rd ⊕ Rd ⊕ Rd ). Then the operator B with the kernel
Z
i
−d
B(x, y) = (2π~)
c(x, p, y)e ~ (x−y)p dp
belongs to Ψ and equals Op(b), where
i~
b(x, p) = e 2 Dp (−Dx +Dy ) c(x, p, y)
.
x=y
Consequently,
b(x, p) = c(x, p, x) +
i~
∂x c(x, p, y) − ∂y c(x, p, y) + O(~2 ).
2
x=y
(5.96)
Proof. We compute:
b(x, p) = (2π~)−d
Z
i
z
z
e ~ z(w−p) c x + , w, x − dzdw,
2
2
then we apply (2.17). 2
5.4
Inequalities
∗
Theorem 5.9. (1) Let b ∈ S and Op(b) = Opa
O(~) in S
∗
∗
,a
∗
(ba
,a
∗
). Then ba
,a
∈ S and b − ba
∗
,a
=
∗
∗
(2) Let ba,a ∈ S and Opa,a (ba,a ) = Op(b). Then b ∈ S and ba,a − b = O(~) in S.
Proof. We use
ba
∗
,a
2
~
2
= e 4 (∂x +∂p ) b,
b=e
2
2
~
4 (∂x +∂p )
2
2
ba,a ,
and the obvious mapping properties of e 4 (∂x +∂p ) . 2
~
33
(5.97)
∗
(5.98)
Theorem 5.10 (Sharp Gaarding Inequality). Let b ∈ S be positive. Then
Op(b) ≥ −C~.
Proof. Let b0 be the Wick symbol of Op(b), that is,
∗
Opa,a (b0 ) = Op(b).
(5.99)
Then, by Thm 5.9 (1), we have b0 ∈ S and b0 − b = O(~) in S. Besides,
∗
Op(b0 ) = Opa
,a
(b).
Now
Op(b) = Op(b0 ) + Op(b − b0 ) = Opa
(5.100)
∗
,a
(b) + O(~).
(5.101)
The first term on the right of (5.101) is positive, because it is the anti-Wick quantization of
a positive symbol. 2
Theorem 5.11 (Fefferman-Phong Inequality). Let b ∈ S be positive. Then
Op~ (b) ≥ −C~2 .
We will not give a complete proof. We will only note that the inequality follows by basic
calculus if we assume that
k
X
c2j
b=
j=1
for real cj ∈ S. We note also that the Sharp Gaarding inequality is true for matrix valued
symbols, with the same proof. This is not the case of the Fefferman-Phong Inequality.
5.5
Semiclassical asymptotics of the dynamics
Theorem 5.12 (Egorov Theorem). Let h be the sum of a polynomial of second order and
a S(Rd ⊕ Rd ) function.
(1) Let x(t), p(t) solve the Hamilton equations with the Hamiltonian h and the initial conditions x(0), p(0). Then
γt (x(0), p(0)) = x(t), p(t)
defines a symplectic (in general, nonlinear) transformation which preserves S(Rd ⊕Rd ).
(2) Let b ∈ S(Rd ⊕ Rd ). Then there exist bt,2j ∈ S(Rd ⊕ Rd ), j = 0, 1, . . . , such that for
|t| ≤ t0
n
X
it
it
e ~ Op(h) Op(b)e− ~ Op(h) −
Op ~2j bt,2j = O(~2n+2 ).
(5.102)
j=0
Moreover,
bt,0 (x, p) = b γt−1 (x, p)
and suppbt,2j ⊂ γt suppb, j = 0, 1, . . . .
34
(5.103)
Proof. We have
n
X
=
j=0
n
X
~2j
it
d − it Op(h)
Op bt,2j e ~ Op(h)
e ~
dt
~2j e
−it
~ Op(h)
j=0
−
it
i
d
Op(h), Op bt,2j + Op
bt,2j e ~ Op(h)
~
dt
(5.104)
(5.105)
This gives us a sequence of equations for bt,2j of the form
where rt,2j
d
bt,0 = {bt,0 , h},
dt
d
bt,2 = {bt,2 , h} + rt,2 ,
dt
......
d
bt,2j = {bt,2j , h} + rt,2j ,
dt
involve bt,0 , . . . , bt,2j−2 . The initial conditions are
(5.106)
(5.107)
(5.108)
(5.109)
bt,0 = b,
(5.110)
bt,2 = 0,
(5.111)
......
(5.112)
bt,2j = 0.
(5.113)
(5.106) is solved by (5.103). (5.109) is solved by
Z t
bt,2j (x, p) =
rt−s,2j γs−1 (x, p) ds.
(5.114)
0
The error term in (5.105) is O(~2n+2 ). Integrating (5.105) from 0 to t we obtain the
estimate (5.102). 2
5.6
Frequency set
Let ~ 7→ ψ~ ∈ L2 (X ). Let (x0 , p0 ) ∈ Rd ⊕ Rd .
Theorem 5.13. The following conditions are equivlent:
(1) There exists b ∈ Cc∞ (Rd ⊕ Rd ) such that b(x0 , p0 ) 6= 0 and
kOp~ (b)ψ~ k = O(~∞ ).
(2) There exists a neighborhood U of (x0 , p0 ) such that for all c ∈ Cc∞ (U)
kOp~ (c)ψ~ k = O(~∞ ).
The set of points in Rd ⊕ Rd that do not satisfy the conditions of Theorem 7.7 is called
the frequency set of ~ 7→ ψ~ and denoted F S(ψ~ ).
Note that we can replace the Weyl quantization by the x, p or p, x quantization in the
definition of the frequency set.
35
5.7
Properties of the frequency set
Theorem 5.14. Let a ∈ S. Then
F S (Op~ (a)ψ~ ) = supp(a) ∩ F S(ψ~ ).
Theorem 5.15. Let h ∈ S + Pol≤2 be real. Let t 7→ γt be the Hamiltonian flow generated
by h. Then
F S(eitOp~ (h) ψ~ ) = γt (F S(ψ~ )) .
Theorem 5.16. Let
i
ψ~ (x) = a(x)e ~ S(x) .
Then
F S(ψ~ ) = {x ∈ suppa, p = ∇S(x)}.
Proof. We apply the nonstationary method. Let p 6= ∂x S(x) on the support of b ∈
d
Cc∞ (Rd ⊕ Rd ). Let (2π~)− 2 F~ denote the unitary semiclassical Fourier transformation.
Then
Z
i
i
p,x
−d
−d
2
2
(2π~) F~ Op~ (b)ψ~ (p) = (2π~)
e− ~ xp b(p, x)a(x)e ~ S(x) dx. (5.115)
Let
T := (p − ∂x S(x))−2 (p − ∂x S(x))∂x .
Let
T # = ∂ˆx (p − ∂x S(x))−2 (p − ∂x S(x))
be the transpose of T . Clearly,
i
i
−i~T e ~ (S(x)−xp) = e ~ (S(x)−xp) .
(5.116)
Therefore, (5.115) equals
d
Z
d
Z
(−i~)n (2π~)− 2
=
(−i~)n (2π~)− 2
i
b(p, x)a(x)T n e ~ (S(x)−xp) dx
i
(5.117)
d
e ~ (S(x)−xp) T #n b(p, x)a(x)dx = O(~n− 2 ).
(5.118)
2
5.8
Algebras with a filtration/gradation
Let S be an algebra.
We say that it is an algebra with filtration {S m : m ∈ Z− } iff S m are linear subspaces
T
0
0
0
of S such that S = S 0 , S m ⊂ S m , m ≤ m0 and S m ·S m ⊂ S m+m . We write S −∞ := S m .
m
Clearly, S −∞ is an ideal and so are S m .
36
0
⊕ S (m) such that S (m) ·S (m ) ⊂
P (m0 )
0
S (m+m ) . Clearly, a gradation induces a filtration by taking S m :=
S
.
We say that S is an algebra with a gradation if S =
m∈Z−
m0 ≤m
Instead of Z+ we can take any ordered group, eg. Z orP
R, then the definition slightly
changes: the full algebra is denoted S ∞ and satisfies S ∞ =
S m . S −∞ is an ideal in S ∞ .
m
S m for m ≥ 0 are ideals in the algebra S 0 .
5.9
Algebra of semiclassical operators
We say that ~ 7→ b~ is a admissible semiclassical symbol if for any n there exist b0 , . . . , bn ∈ S
and ~ 7→ r~ ∈ S such that for any n
b=
n
X
~j bj + ~n+1 r~
j=0
α β k ∂x ∂p ∂~ r~ ≤ Cα,β,k .
Note that the sequence b0 , b1 , . . . is uniquely defined (does not depend on n).
Let Θ ⊂ Rd ⊕ Rd be closed. We say that b~ is O(~∞ ) outside Θ if b0 , b1 , · · · = 0 outside
Θ.
Let Ssc denote the space of admissible semiclassical symbols and Ψsc the set of their
semiclassical quantizations. We write Ssc (Θ) for the space of symbols that vanish outside Θ
and Ψsc (Θ) for their quantizations.
Clearly, Ψsc is an algebra with gradation given by Ψ−m
:= ~m Ψsc . It is closed wrt
sc
operations described in Theorems. Ψsc (Θ) are ideals in Ψsc .
5.10
Principal and subprincipal symbols
P∞
Recall that by s(A~ ) = a~ (x, ξ) we denote the symbol of A~ . We have a~ ' m=0 ~j a−j .
We call a0 the principal and a−1 the subprincipal symbol. Let us introduce the extended
principal sumbol:
sep (A) := a0 + ~a−1 .
P∞ k
x,p
If A = Op (b) and b ' k=0 ~ b−k , then the principal symbol is b0 and the subprincipal
symbol is b−1 + i~
2 ∂x ∂ξ b0 .
Theorem 5.17. Let A ∈ Ψsc and A0 ∈ Ψsc . Then
A~ A0~ ∈ Ψsc
1
sep
[A~ , A0~ ]+
2
[A~ , A0~ ] ∈ ~Ψsc
sep ~−1 [A~ , A0~ ]
and
=
(5.119)
sep (A~ )sep (A0~ ) (mod ~2 ),
(5.120)
{sep (A~ ), sep (A0~ )} (mod ~2 ).
(5.122)
and
=
(5.121)
37
5.11
Algebra of formal semiclassical operators
We also have the space of formal semiclassical symbols, which are formal power series
b=
n
X
~j bj
j=0
f
with coefficients, say, in S. We denote this space by Ssc
and their quantizations by Ψfsc .
Theorem 5.18. We have the isomorphism
Ψsc /Ψ−∞
= Ψfsc .
sc
5.12
More about star product
The star product can be written in the following asymmetric form:
b ∗ c(x, p)
=
=
~
b x − Dp , p +
2
~
b x − Dp , p +
2
~ Dx c(x, p)
2
~ Dx c(x, p).
2
(5.123)
(5.124)
Note that the operators x − ~2 Dp and p + ~2 Dx commute. Thus we can understand the rhs
of(5.123) as a function of two commuting operators. Alternatively, we can understand it as
the PDE notation for the x, p quantization. Alternatively, (5.124) gives us an interpretation
in terms of the Weyl quantization.
To see this we start from
i
b ∗ c(x, p) = e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) b(x1 , p1 )c(x2 , p2 ) x := x = x , .
(5.125)
1
2
p := p1 = p2 .
i
We treat b(x, p) as the operator of multiplication. We move e 2 ~(Dp1 Dx2 −Dx1 Dp2 ) to the right
obtaining
~
~
b x1 − Dp2 , p1 + Dx2 c(x2 , p2 ) x := x = x , ,
1
2
2
2
p := p1 = p2 .
which equals the LHS of (5.125).
38
Note the following consequences:
f Op(h) Op(b)
! !
~
~ = Op f h x − Dp , p + Dx b ,
2
2
f Op(h)
=
Op f
! !
1 ~
~ 1 ~
~ h x − Dp , p + Dx + h x + Dp , p − Dx 1 ,
2
2
2
2
2
2
i
i
e ~ Op(h) Op(b)e− ~ Op(h)
! !
i
~ ~ ~
~
= Op exp
h x − Dp , p + Dx − h x + Dp , p − Dx b .
~
2
2
2
2
We have similar formulas for the product in the x, p-quantization. Let
Opx,p (a) = Opx,p (b)Opx,p (c).
Then
a(x, p)
6
6.1
= b x, p + ~Dx c(x, p)
= c x + ~Dp , p b(x, p).
Spectral asymptotics
Functional calculus and Weyl asymptotics
It follows from (3.44) that
Op(b)n = Op(bn ) + O(~2 ).
Hence for polynomial functions
f Op(b) = Op(f ◦ b) + O(~2 ).
(6.126)
One can expect (6.126) to be true for a larger class of nice functions. Consequently,
Trf Op(b) = Tr Op(f ◦ b) + O(~2 )
Z
= (2π~)−d f b(x, p) dxdp + O(~−d+2 ).
(6.127)
In particular, we can try to use f = 1l]−∞,µ] . It is too optimistic to expect
1l]−∞,µ] (Op(h)) = Op 1l]−∞,µ] (h) + O(~2 ).
(6.128)
After all the step function is not nice – it is not even continuous. If there is a gap in the
spectrum around µ, one can try to smooth it out. Therefore, there is a hope for some weaker
error term instead of O(~2 ).
39
6.2
Weyl asymptotics
For a bounded from below self-adjoint operator H set
Nµ (H) := #{eigenvalues of H counted with multiplicity ≤ µ}
= Tr1l]−∞,µ] (H).
(6.129)
(6.130)
If (6.128) were true, then we would have
Nµ Op(h)
=
(2π~)−d
Z
dxdp + O(~−d+2 ).
(6.131)
h(x,p)≤µ
Asymptotics of the form (6.131) are called the Weyl asymptotics. In practice the error
term O(~−d+2 ) may be too optimistic and one gets something worse (but hopefully at least
o(~−d )).
We will show that if V is continuous potential with V − µ > 0 outside a compact set
then
Z
d
(6.132)
Nµ (−~2 ∆ + V (x)) ' (2π~)−d cd
|V (x) − µ|−2 dx + o(~−d ).
V (x)≤µ
This is (6.131) for H = p2 + V (x) with a rather weak error estimate.
We will prove this without using the Weyl calculus. Here are the tools that we will use:
A ≤ B ⇒ Nµ (A) ≥ Nµ (B),
Nµ (A ⊕ B) = Nµ (A) + Nµ (B).
To simplify we will assume that d = 1.
Lemma 6.1. Let ∆[0,L],D , resp. ∆[0,L],N denote the Dirichlet, resp. Neumann Laplacian
on [0, L]. For α ∈ R let [α] denote the largest integer ≤ α, θ(α) the Heavyside function and
|µ|+ := µθ(µ). Then
1/2
Nµ − ~2 ∆[0,L],D
=
[L(π~)−1 |µ|+ ],
Nµ − ~2 ∆[0,L],N
=
[L(π~)−1 |µ|+ ] + θ(µ).
1/2
Proof. The eigenfunctions and the spectrum of ∆[0,L],D , resp. ∆[0,L],N are
πnx
,
L
πnx
cos
,
L
sin
~2 π 2 n2
,
L2
~2 π 2 n2
,
L2
n = 1, 2, . . . ;
n = 0, 1, 2, . . . .
1/2
Thus the last eigenvalue has the number n = [L(~π)−1 |µ|+ ]. 2
Divide R into intervals
h
i
IN,j := (j − 1/2)N −1 , (j + 1/2)N −1 .
40
Put at the borders of the intervals Neumann/Dirichlet boundary conditions. The Neumann
conditions lower the expectation value and the Dirichlet conditions increase them. Set
V N,j
=
sup{V (x) : x ∈ IN,j },
V N,j
=
inf{V (x) : x ∈ IN,j }.
We have
⊕
− ~2 ∆IN,j ,N + V N,j
− ~2 ∆IN,j ,D + V N,j .
j∈Z
≤ −~2 ∆ + V (x) ≤
⊕
j∈Z
Hence,
X
Nµ − ~2 ∆IN,j ,N + V N,j
j∈Z
≥ Nµ
− ~2 ∆ + V (x)
≥
X
Nµ − ~2 ∆IN,j ,D + V N,j .
j∈Z
Therefore,
X
1/2
N −1 (~π)−1 |V N,j − µ|− +
≥ Nµ − ~2 ∆ + V (x)
≥
X
θ µ − V N,j
j∈Z
j∈Z
X
N −1 (~π)−1 |V N,j −
1/2
µ|− .
j∈Z
Using the fact that |V − µ|− has a compact support, we can estimate
X
θ µ − V N,j
≤ N C.
j∈Z
By properties of Riemann sums we can find N such that for N ≥ N
Z
X
1/2
1/2
−1
N |V N,j − µ|− − |V (x) − µ|− dx < /3,
(6.133)
j∈Z
Z
X
1/2
1/2
−1
N |V N,j − µ|− − |V (x) − µ|− dx <
/3.
(6.134)
j∈Z
Therefore,
Z
1
1/2
2
|V (x) − µ|− dx
Nµ − ~ ∆ + V (x) −
~π
<
2
CN
+
.
~π3
π
(6.135)
Hence the right hand side of (6.135) is o(~−1 ). This proves (6.132)
If we assume that V is
to be C0 −1 . Then we
√ differentiable, then N can be assumed
−1
can optimize and set = ~. This allows us to replace o(~ ) by O(~−1/2 ).
41
6.3
Energy of many fermion systems
Consider fermions with the 1-particle space is spanned by an orthonormal basis Φ1 , Φ2 , . . . .
The n-particle fermionic space is spanned by Slater determinants
1
Ψi1 ,...,in := √ Φi1 ∧ · · · ∧ Φin ,
n!
i1 < · · · < in .
Suppose that we have noninteracting fermions with the 1-particle Hamiltonian H. Then
the Hamiltonian on the N -particle space will be denoted by dΓn (H).
Suppose that E1 < E2 < . . . are the eigenvalues of H in the ascending order and
Φ1 , Φ2 , . . . are the corresponding normalized eigenvectors. This means that the full Hamiltonian dΓn (H) acts on Slater determinants as
dΓn (H)Ψi1 ,...,in = (Ei1 + · · · + Ei1 )Ψi1 ,...,in .
For simplicity we assume that eigenvalues are nondegenerate. Then the ground state of
the system is the Slater determinant
1
Ψ := √ Φ1 ∧ · · · ∧ Φn .
n!
(6.136)
The ground state energy is E1 + · · · + En .
In practice it is often more convenient as the basic parameter to use the chemical potential
µ instead of the number of particles n. Then the 1-particle density matrix of the ground
state is given by 1l]−∞,µ] (H), where we find µ from the relation
Tr1l]−∞,µ] (H) = n.
If B is a 1-particle observable, then its expectation value in an n-fermionic state Ψ is
given by
ΨdΓn (B)Ψ = TrBγΨ
where γΨ is the so-called reduced 1-particle density matrix. Note that 0 ≤ γΨ ≤ 1l and
TrγΨ = n. The reduced 1-particle density matrix of the Slater determinant (6.136) Ψ is the
projection onto the space spanned by Φ1 , . . . , Φn .
Suppose that the 1-particle space is L2 (Rd ). Then the 1-particle reduced density matrix
can be represented by its kernel γΨ (x, y). Expliccitly,
Z
Z
(6.137)
γΨ (x, y) = dx2 · · · dxn Ψ(x, x2 , · · · , xn )Ψ(y, x2 , · · · , xn ).
We are particularly interested in expectation values of the position. For position independent observables we do not need to know the full reduced density matrix, but only the
density:
Z
TrγΨ f (x̂) = ρΨ (x)f (x),
42
where
ρΨ (x) := γΨ (x, x).
Note that
Z
ρΨ (x)dx = n.
(6.138)
If γ = Op(g), then
TrγΨ f (x̂) = (2π~)−d
Z Z
g(x, p)f (x)dxdp.
Hence
ρΨ (x) = (2π~)−d g(x, p)dp.
Suppose now that the 1-particle Hamiltonian is H = Op(h). Remember that then the
symbol of 1l]−∞,µ] (H) is approximately given by
1l]−∞,µ] h(x, p) .
Hence, the reduced 1-particle density matrix of the ground state is γ = 1l]−∞,µ] (H). The
corresponding density is
Z
Z
−d
−d
ρ(x) ≈ (2π~)
1l]−∞,µ] h(x, p) dp = (2π~)
dp.
h(x,p)≤µ
Let cd rd be the volume of the ball of radius r. If h(x, p) = p2 + v(x), then
Z
ρ(x) ≈ (2π~)−d
dp
p2 +v(x)≤µ
d
=
(2π~)−d cd |v(x) − µ|−2 .
Let us compute the kinetic energy
Z Z
−d
2
Trp̂ 1l]−∞,µ] (H) ≈ (2π~)
p2 dxdp
p2 +v(x)<µ
=
−d
Z
(2π~)
Z
dx
dcd |p|d+1 d|p|
1
2
|p|<|v(x)−µ|−
Z
d+2
dcd
|v(x) − µ|−2
d+2
Z
d+2
d
−2/d
−d
≈ (2π~)
c
ρ d (x)dx.
d+2 d
=
(2π~)−d
dx
Thus if we know that ρ is the density of a ground state of a Schrödinger Hamiltonian, then
we expect that the kinetic energy is given by the functional
Z
d+2
d −2/d
Ekin (ρ) := (2π~)−d
cd
ρ d (x)dx.
(6.139)
d+2
43
Clearly, the potential energy of a state with density ρ in the potential V is given by
Z
Epot (ρ) := V (x)ρ(x)dx.
(6.140)
Suppose in addition that the fermions interact with the potential W . That is, we consider
the Hamiltonian
N
X
X
p2i + V (xi ) +
W (xi − xj )
(6.141)
i=1
1≤i<j≤N
on the N -particle antisymmetric space ∧N L2 (Rd ).
Then we can expect by classical arguments that for a state Ψ
Z Z
X
Ψ|
W (xi − xj )Ψ '
W (x − y)ρΨ (x)ρΨ (y)dxdy.
1≤i<j≤N
This suggests to introduce the interaction functiona
Z Z
Eint (ρ) :=
W (x − y)ρ(x)ρ(y)dxdy.
(6.142)
The Thomas-Fermi functional is given by the sum of (6.139), (6.140) and ((6.142):
ETF (ρ) := Ekin (ρ) + Epot (ρ) + Eint (ρ)
Z
d+2
d −2/d
cd
ρ d (x)dx
= (2π~)−d
d+2
Z
Z Z
+ V (x)ρ(x)dx +
W (x − y)ρ(x)ρ(y)dxdy.
(6.143)
(6.144)
We expect that
Z
n
inf ETF (ρ) | ρ ≥ 0,
ρ(x)dx = n
o
(6.145)
approximates the ground state energy of (6.141).
7
Standard pseudodifferential calculus on Rd
7.1
Classes of symbols
We will often write X = Rd .
m
Let m ∈ N. We define Spol
(T# Rd ) to be the set of functions of the form
a(x, ξ) =
X
aβ (x)ξ β ,
(7.146)
|β|≤m
where
|∂xα aβ | ≤ cα,β .
44
(7.147)
Let m ∈ R. We define S m (T# Rd ) to be the set of functions a ∈ C ∞ (T# Rd ) such that
for any
|∂xα ∂ξβ a(x, ξ)| ≤ cα,β hξim−|β| .
(7.148)
m
We set Sph
(T# Rd ) to be the set of functions a ∈ S m (T# Rd ) such that for any n there
exist functions am−k , k = 0, . . . , n, homogeneous of degree m − k such that
|∂xα ∂ξβ am−k (x, ξ)| ≤
n
X
α β
am−k (x, ξ) ≤
∂x ∂ξ a(x, ξ) −
cα,β |ξ|m−k−|β| , |ξ| > 1,
cα,β,n |ξ|m−k−n−1 . |ξ| > 1.
k=0
P∞
We then write a ' k=0 am−k , where am−k are uniquely determined.
We introduce also
m
S −∞ := ∩ S m = ∩ Sph
,
m
S
Clearly, for m ∈ N,
7.2
∞
m
:= ∪ S ,
m
Spol
m
⊂
m
Sph
.
∞
Sph
m
:=
m
∪ Sph
,
m
For any m ∈ R,
∞
m
Spol
:= ∪ Spol
.
m
m
Sph
m
⊂S .
Classes of pseudodifferential operators
m
Let A be an operator from S(Rd ) to S 0 (Rd ). We will say that A ∈ Ψm , Ψm
ph , Ψpol iff
A = Op(a)
m
m
for a ∈ S m , Sph
, Spol
. (Note, that instead of the Weyl quantization we could use the KohnNirenberg quantization as well, obtaining the same class of operators).
For k ∈ R, the kth Sobolev space is defined as
L2,k (Rd ) := (1 − ∆)−k/2 L2 (Rd ).
We also write
L2,∞ := ∩ L2,k (Rd ),
k
L2,−∞ := ∪ L2,k (Rd ).
k
The Calderon-Vaillancourt Theorem implies immediately the following theorem:
Theorem 7.1. For any k, m ∈ R, A ∈ Ψm extends to a bounded operator
A : L2,k (Rd ) → L2,k−m (Rd ),
and also to a continuous operator on L2,∞ and L2,−∞ . A ∈ Ψ−∞ maps L2,−∞ to L2,∞ .
45
7.3
Extended principal symbol
(m−2)
We will write sep (A) for s(A)
) and call it the extended principal symbol of a.
P∞(modS
If A = Op(a) and a ' k=0 am−k , then am is called the principal symbol and am−1 the
subprincipal symbol of a. We
then identify aep with am + am−1 .
Pcan
∞
If A = Opx,p (b) and b ' k=0 bm−k , then the principal symbol is bm and the subprincipal
symbol is bm−1 + 2i ∂x ∂ξ bm .
0
Theorem 7.2. Let A ∈ Ψm and A0 ∈ Ψm . Then
0
AA0 ∈ Ψm+m
1
sep [A, A0 ]+
2
0
0
[A, A ] ∈ Ψm+m −1
0
sep ([A, A ])
7.4
and
=
(7.149)
0
sep (A)sep (A0 )(mod S m+m −2 )
(7.150)
and
=
(7.151)
0
{sep (A), sep (A )}(mod S
m+m0 −3
).
(7.152)
Diffeomorphism invariance
Let F : Rd → Rd be a diffeomorphism that moves only a bounded part of Rd . We can define
its prolongation to the cotangent bundle T# Rd , denoted T# F and the corresponding acttion
on S(Rd ), denoted F# .
m
m
Theorem 7.3. The spaces Spol
(T# Rd ), Sph
(T# Rd ), S m (T# Rd ) are invariant wrt T# F .
The operators F# are bounded invertible on spaces L2,m .
The algebras Ψpol (Rd ), Ψph (Rd ) and Ψ(Rd ) are invariant wrt F# .
7.5
Ellipticity
Let m ≥ 0. We say that b ∈ S m (Rd ) is elliptic if b is real and for some r, c0 > 0
b(x, ξ) ≥ c0 |ξ|m , |ξ| > r.
Theorem 7.4. Let p ∈ R. Let b ∈ S m (Rd ) be elliptic and z − b(x, ξ) invertible.
(1) (−z + b(x, ξ))p is well defined and belongs to S mp (T # Rd ).
(2) Op(b) is essentially selfadjoint and there exists c such that for z ∈ C\[c, ∞[
z 6∈ sp Op(b), (z − Op(b))m ∈ Ψmp (Rd ).
7.6
Asymptotics of the dynamics
1
Theorem 7.5 (Egorov Theorem). Let h ∈ Sph
be real and elliptic.
(1) Let x(t), ξ(t) solve the Hamilton equations with the Hamiltonian h and the initial conditions x(0), ξ(0). Then
γt (x(0), ξ(0)) = x(t), ξ(t)
defines a symplectic transformation homogeneous in ξ.
46
m
(2) Let b ∈ Sph
. Then there exist bt,m−2j ∈ S m−2j , j = 0, 1, . . . , and Rt,m−2n−2 uniformly
in B(L2,k , L2,k−2n−2 )). such that for |t| ≤ t0
eitOp(h) Op(b)e−itOp(h) =
n
X
Op bt,m−2j + Rt,m−2n−2 .
(7.153)
j=0
Moreover,
bt,m (x, ξ) = b γt (x, ξ)
(7.154)
and suppbt,m−2j ⊂ γt−1 suppb, j = 0, 1, . . . .
We will say that a ∈ S m has a homogeneous principal symbol if there exists am ∈ S m
homogeneous of degree m in ξ and m1 < m such that a − am ∈ S m1 for |ξ| > 1.
7.7
Singular support
Let f ∈ Dc0 (Rd ). It belongs to Cc∞ (Rd ) iff
|fˆ(ξ)| ≤ cn hξi−n ,
n ∈ N.
(7.155)
Let f ∈ D0 (Rd ) and x0 ∈ Rd . Clearly, f is C ∞ in a neighborhood of x0 iff
c (ξ)| ≤ cn hξi−n ,
∃χ∈Cc∞ (Rd ) χ(x0 ) 6= 0, |χf
n ∈ N..
(7.156)
This is equivalent to saying that
∃U a neighborhood of
∞
∀
x0 χ∈Cc (U )
c (ξ)| ≤ cn hξi−n ,
|χf
n ∈ N.
(7.157)
The complement of points where f is C ∞ is called the singular support of f .
7.8
Preservation of the singular support
Theorem 7.6. Let f ∈ L2,−∞ and A ∈ Ψ∞ . Then Af is C ∞ away from suppf .
Proof. Let f ∈ L2,k and A ∈ Ψm . Let χ, χ1 , ∈ C ∞ have all bounded derivatives. Let χ = 1
on suppf and χ1 = 0 on suppχ. Let
χ1 (x)Af = χ1 (x)Aχ(x)n f = χ1 (x)adnχ (A)
(7.158)
But adnχ (A) ∈ Ψm−n . Thus (7.158) belongs to L2,k−m+n . 2
7.9
Wave front
We equip X ⊕ (X # \{0}) with an action of R+ as follows:
(x, ξ) 7→ (x, tξ).
47
(7.159)
We say that a subset of X ⊕ (X # \{0}) is conical iff it is invariant with respect to this action.
Conical subsets can be identified with X ⊕ (X # \{0})/R+ .
Let (x0 , ξ0 ∈ X ⊕ X # \{0}. We say that f is smooth in a conical neighborhood of (x0 , ξ0 )
iff there exists χ ∈ Cc∞ (X ) and a conical neighborhood U of ξ ∈ U such that
ˆ (ξ)| ≤ cn hξi−n , n ∈ N.
|χf
(7.160)
The complement in X ⊕ (X # \{0})/R+ . of points where f is smooth is called the wave
front set of f and denoted W F (f ).
∞
P
m
m
We will write Sph
(Θ) for the set of a ∈ Sph
with a '
am−k such that suppam−k ⊂ Θ.
k=0
The following theorem gives two possible alternative definitions of the wave front set.
Theorem 7.7. Let u ∈ L−∞ . The following conditions are equivalent:
m
(1) There exists b ∈ Sph
such that bm (x0 , ξ0 ) 6= 0 and
Op(b)u ∈ L2,∞ .
m
(2) There exists a conical neighborhood U of (x0 , ξ0 ) such that for all c ∈ Sph
(U)
Op(c)u ∈ L2,∞ .
7.10
Properties of the wave front set
Example 7.8. Let Y be a k-dimensional submanifold with a k-form β. Then the distribution
Z
hF |ψi :=
φβ
Y
has the wave front set in the conormal bundle to Y:
W F (F ) ⊂ N # Y := {(x, ξ) : x ∈ Y, hξ|vi = 0, v ∈ TY}.
Example 7.9. For X = R,
W F ((x + i0)−1 ) = {(0, ξ) : ξ > 0}.
Example 7.10. Let H be a homogeneous function of degree 1 smooth away from the origin
and v ∈ C ∞ ,
|∂ξβ v(ξ)| ≤ cβ hξim−|β|
Then
Z
eixξ−iH(ξ) v(ξ)dξ = u(x)
satisfies
W F (u) = {(∇ξ H(ξ), ξ) : ξ ∈ suppv}.
48
Theorem 7.11. Let u ∈ L2,−∞ and a '
∞
P
k=0
m
am−k ∈ Sph
. Then
∞
W F (Op(a)u) = W F (u) ∩ ∩ suppam−k .
k=0
1
Theorem 7.12 (Theorem about propagation of singularities). Let h ∈ Sph
be real and
elliptic. Then
W F (eitOp(h) u) = γt (W F (u)) .
8
Path integrals
In this section ~ = 1 and we do not put hats on p and x. We will be not very precise
concerning the limits – often lim may mean the strong limit.
8.1
Evolution
Suppose that we have a family of operators t 7→ B(t) depending on a real variable. Typically,
we will assume that B(t) are generators of 1-parameter groups (eg. i times a self-adjoint
operator). Under certain conditions on the continuity that we will not discuss there exists
a unique operator function that in appropriate sense satisfies
d
U (t+ , t− )
dt+
U (t, t)
= B(t+ )U (t+ , t− ),
=
1l.
It also satisfies
d
U (t+ , t− ) =
dt−
U (t2 , t1 )U (t1 , t0 ) =
−U (t+ , t− )B(t− ),
U (t2 , t0 ).
If B(t) are bounded then
U (t+ , t− ) =
∞
X
Z
Z
B(tn ) · · · B(t1 )dtn · · · dt1 .
...
n=0 t >t >···>t >t
+
n
1
−
We will write
Z
!
t+
Texp
B(t)dt
:= U (t+ , t− ).
t−
In particular, if B(t) = B does not depend on time, then U (t+ , t− ) = e(t+ −t− )B .
In what follows we will restrict ourselves to the case t− = 0 and t+ = t and we will
consider
Z
t
U (t) := Texp
B(s)ds .
0
49
(8.161)
Note that the whole evolution can be retrieved from (8.161) by
U (t+ , t− ) = U (t+ )U (t− )−1 .
We have
n
Y
U (t) = lim
n→∞
t
jt
e n B( n ) .
(8.162)
j=1
(In multiple products we will assume that the factors are ordered from the right to the left).
Now suppose that F (s, u) is an operator function such that uniformly in s
euB(s) − F (s, u)
=
o(u),
kF (s, u)k
≤
C.
Then
U (t) = lim
n
Y
n→∞
F
j=1
jt t ,
.
n n
(8.163)
Indeed,
n
Y
e
jt
t
n B( n )
−
j=1
=
F
j=1
n
n
X
Y
k=1 j=k+1
= no(n−1 )
n
Y
jt t ,
n n
!
kt t k−1
jt t t kt
Y t jt
F
,
e n B( n ) − F
,
e n B( n )
n n
n n
j=1
→
0.
n→∞
Example 8.1. (1) F (s, u) = 1l + uB(s). Thus
U (t) = lim
!
t jt 1l + B
.
n
n
n
Y
n→∞
j=1
Strictly speaking, this works only if B(t) is uniformly bounded.
In particular,
etB = lim
n→∞
1l +
t n
B .
n
−1
(2) F (s, u) = 1l − uB(s)
. Then
U (t) := lim
n→∞
n
Y
t jt 1l − B
n
n
j=1
This should work also if B(t) is unbounded.
In particular,
etB = lim
n→∞
50
1l −
t −n
B
.
n
!−1
.
(3) Suppose that B(t) = A(t)+C(t), where both A(t) and C(t) are generators of semigroups.
Set F (s, u) = euA(t) euC(t) . Thus
U (t) = lim
n
Y
n→∞
jt
t
t
jt
e n A( n ) e n C( n ) .
(8.164)
j=1
In particular, we obtain the Lie-Trotter formula
t t n
et(A+C) = lim e n A e n C .
n→∞
8.2
Scattering operator
We will usually assume that the dynamics is generated by iH(t) where H(t) is a self-adjoint
operator. Often,
H(t) = H0 + V (t),
where H0 is a fixed self-adjoint operator. The evolution in the interaction picture is
Z t+
H(t)dt e−it− H0 .
S(t+ , t− ) := eit+ H0 Texp − i
t−
The scattering operator is defined as
S
:=
lim
t+ ,−t− →∞
S(t+ , t− ).
Introduce the Hamiltonian in the interaction picture
HInt (t) := eitH0 V (t)e−itH0 .
Note that
∂t+ S(t+ , t− )
=
−iHInt (t+ )S(t+ , t− ),
∂t− S(t+ , t− )
=
iS(t+ , t− )HInt (t+ ),
S(t, t)
=
1l.
Therefore,
S(t+ , t− )
=
Z
Texp − i
=
Z
Texp − i
t+
HInt (t)dt ,
t−
∞
S
HInt (t)dt
−∞
8.3
Bound state energy
Suppose that Φ0 and E0 , resp. Φ and E are eigenvectors and eigenvalues of H0 , resp H, so
that
H0 Φ0 = E0 Φ0 , HΦ = EΦ.
51
We assume that Φ, E are small perturbations of Φ0 , E0 when the coupling constant λ is
small enough.
The following heuristic formulas can be sometimes rigorously proven:
E − E0
=
lim (2i)−1
t→±∞
d
log(Φ0 |e−itH0 ei2tH e−itH0 Φ0 ).
dt
(8.165)
To see why we can expect (8.165) to be true, we write
(Φ0 |e−itH0 ei2tH e−itH0 Φ0 ) = |(Φ0 |Φ)|2 ei2t(E−E0 ) + C(t).
Then, if we can argue that for large t the term C(t) does not play a role, we obtain (8.165).
8.4
Path integrals for Schrödinger operators
We consider
1 2
p + V (t, x),
2
1
H(t) := Op(h(t)) = − ∆ + V (t, x),
2
Z t
U (t) := Texp −i
H(s)ds .
h(t, x, p)
:=
(8.166)
0
We have
2
i
i
e− 2 t∆ (x, y) = (2πit)−d/2 e 2t (x−y) .
From
U (t)
=
lim
n→∞
n
Y
t
jt
t
e−i n V ( n ,x) ei 2n ∆
j=1
we obtain
Z
U (t, x, y)
=
=
n→∞
lim
n Y
2πit − d2
Z
dxn−1 · · ·
lim
dx1
n
j=1
Z
2πit − dn
2
n→∞
× exp
n
e
in(xj−1 −xj )2
2t
Z
dxn−1 · · ·
dx1
!
n
jt
it X n2 (xj−1 − xj )2
−V
, xj
y=x , .
0
n j=1
2t2
n
x = xn .
Heuristically, this is written as
Z
Z t
L s, x(s), ẋ(s) ds Dx,y x(·) ,
U (t, x, y) = exp i
0
52
y=x ,
0
x = xn .
t
−i n
V ( jt
n ,xj )
where
L(s, x, ẋ) :=
1 2
ẋ − V (s, x)
2
is the Lagrangian and
2πit − dn
t
(n − 1)t 2
· · · dx
Dx,y (x(·) := lim
dx
n→∞
n
n
n
(8.167)
is some kind of a limit of the Lebesgue measure on paths [0, t] 3 s 7→ x(s) such that x(0) = y
and end up at x(t) = x.
8.5
Example–the harmonic oscillator
Let
H=
1 2 1 2
p + x .
2
2
It is well-known that for t ∈]0, π[,
1
e−itH (x, y) = (2πi sin t)− 2 exp
−(x2 + y 2 ) cos t + 2xy 2i sin t
.
(8.168)
(8.168) is called the Mehler formula.
We will derive (8.168) from the path integral formalism. We will use the explicit formula
for the free dynamics with H0 = 12 p2 :
1
e−itH0 (x, y) = (2πit)− 2 exp
−(x − y)2 2it
.
(8.169)
For t ∈]0, π[, there exists a unique trajectory for H starting from y and ending at x. Similarly
(with no restriction on time) there exists a unique trajectory for H0 :
cos(s − 2t )
sin(s − 2t )
(x
+
y)
+
(x − y),
cos 2t
sin 2t
(t − s)
s
.
x0,cl (s) = x + y
t
t
xcl (s) =
(8.170)
(8.171)
Now we set x(s) = xcl (s) + z(s) and obtain
Z
0
t
Z
t
1 2
ẋ (s) − x2 (s) ds
2
0
Z t
Z t
=
L xcl (s), ẋcl (s) ds +
L z(s), ż(s) ds
0
0
Z t
(x2 + y 2 ) cos t − 2xy
1 2
=
+
ż (s) − z 2 (s) ds.
2 sin t
0 2
L x(s), ẋ(s) ds =
53
(8.172)
(8.173)
(8.174)
Similarly, setting x(s) = x0,cl (s) + z(s) we obtain
Z
t
L0 x(s), ẋ(s) ds =
0
Z
t
0
1 2
ẋ (s)ds
2
(8.175)
(8.176)
2
=
(x − y)
+
2t
t
Z
0
1 2
ż (s)ds.
2
(8.177)
Therefore,
R
t
exp
i
L
x(s),
ẋ(s)
ds Dx,y x(·)
0
e
(x, y)
R
=R
t
e−itH0 (x, y)
exp i 0 L0 x(s), ẋ(s) ds Dx,y x(·)
2 2
R
Rt
cos t−2xy
exp i (x +y 2)sin
+ i 0 12 ż 2 (s) − z 2 (s) ds D0,0 z(·)
t
=
R
Rt 1
2
2 (s)ds D
exp i (x−y)
+
i
ż
0,0 z(·)
2t
0 2
2 2
! 21
(x +y ) cos t−2xy
i
exp
i
(−∆)
2 sin t
= det i 2
2
(x−y)
exp i
2 (−∆ − 1)
−itH
R
(8.178)
(8.179)
(8.180)
2t
Here −∆ denotes the minus
o the Dirichlet boundary conditions on the interval
n 2 2 Laplacian with
[0, t]. Its spectrum is π t2k | k = 1, 2, . . . . Therefore, at least formally,
det
i
2 (−∆)
i
2 (−∆
− 1)
!
1
=
det 1l + ∆−1
=
∞ Y
1−
k=1
(8.181)
t2 t
=
.
2
2
π k
sin t
(8.182)
Now (8.169) implies (8.168).
8.6
Path integrals for Schrödinger operators with the imaginary
time
Let us repeat the same computation for the evolution generated by
−H(t) = − − ∆ + V (t, x) .
We add the superscript E for “Euclidean”:
Z t
U E (t) := Texp −
H(s)ds
=
0
54
lim
n→∞
n
Y
j=1
t
jt
t
e− n V ( n ,x) e n ∆ .
Using
2
1
1
e 2 t∆ (x, y) = (2πt)−d/2 e− 2t (x−y) .
we obtain
E
U (t, x, y)
=
Z
2πt − dn
2
lim
n→∞
× exp
Z
dxn−1 · · ·
n
dx1
!
n
jt
t X −n2 (xj − xj−1 )2
−V
, xj
y=x , .
0
n j=1
2t2
n
x = xn .
Heuristically, this is written as
Z
Z t
E
E
LE s, x(s), ẋ(s) ds Dx,y
x(·) ,
U (t, x, y) = exp −
0
where
LE (s, x, ẋ) :=
1 2
ẋ + V s, x
2
is the “Euclidean Lagrangian” and
2πt − dn
(n − 1)t t
2
E
Dx,y
(x(·) := lim
dx
· · · dx
n→∞
n
n
n
is similar to (8.167).
8.7
Wiener measure
dWy x(·) = exp −
Z
t
0
1 2 E
ẋ (s) dsDx(t),y
x(·) dx(t)
2
can be interpreted as a measure on paths, functions [0, t] 3 s 7→ x(s) such that x(0) = y—the
Wiener measure.
Let us fix tn > · · · > t1 > 0, and F is a function on the space of paths depending only
on x(tn ), . . . , x(t1 ) (such a function is called a cylinder function). Thus
F x(·) = Ftn ,...,t1 x(tn ), . . . , x(t1 ) .
Then we set
Z
dWy x(·) F x(·)
Z
=
Ftn ,...,t1 (xn , . . . , x1
(8.183)
e
−
(xn −xn−1 )2
2(tn −tn−1 )
e
−
(x2 −x1 )2
2(t2 −t1 )
e−
(x1 −y)2
2t1
d dxn · · ·
d dx2
d dx1 .
2π(tn − tn−1 ) 2
2π(t2 − t1 ) 2
2πt1 2
We easily check the correctness of the definition on all cylinder functions. Then we extend
the measure to a larger space of paths–there are various possibilities.
55
We can use the Wiener measure to (rigorously) express the integral kernel of U E (t). Let
Φ, Ψ ∈ L2 (Rd ). Then the so-called Feynman-Katz formula says
(Φ|U E (t)Ψ)
Z
Z
Z t
V (s, x(s) ds .
= dx(0) dWx(0) x(·) Φ x(t) Ψ x(0) exp −
(8.184)
0
Theorem 8.2. Let t, t1 , t2 > 0. Then
Z
x(t)dW0 x(·) = 0,
Z
xi (t2 )xj (t1 )dW0 x(·) = δij min(t2 , t1 ),
Z
2
x(t2 ) − x(t1 ) dW0 x(·) = d|t2 − t1 |.
(8.185)
(8.186)
(8.187)
Proof. Let us prove (8.186). Let t2 > t1 . Then
Z
x(t2 )x(t1 )dW0 x(·) =
Z Z
x2
Z
=
e
−
x2
1
(x2 −x1 )2
2(t2 −t1 )
d
(2π(t2 − t1 )) 2
x1
e− 2t1
d
(2πt1 ) 2
dx1 dx2
(8.188)
x2
1
x21
e− 2t1
dx1
d
(2πt1 ) 2
=
t1 .
(8.189)
2
Recall the formula (2.16)
e
1
2 ∂x ·ν∂x
Ψ(0)
=
− 21
(det 2πν)
Z
1
Ψ(x)e− 2 x·ν
−1
x
dx,
(8.190)
which says that for Gaussian measures you can “integrate by differentiating”. The Wiener
measure is Gaussian, and in this case (8.190) has the form
Z
1
dW0 x(·) F x(·) = exp
∂x(s2 ) min(s2 , s1 )∂x(s1 ) F x(·) .
(8.191)
2
Indeed, the operator whose quadratic form appears in the Wiener measure is the Laplacian
on [0, t], which is Dirichlet at 0 and Neumann at t. Now the operator with the integral
kernel min(t2 , t1 ) is the inverse of this Laplacian.
8.8
General Hamiltonians – Weyl quantization
Let [0, t] 3 s 7→ h(s, x, p) ∈ R be a time dependent classical Hamiltonian. Set
H(s) := Op h(s)
and U (t) as in (8.201).
56
Lemma 8.3.
e−iuOp
h(s)
− Op e−iuh(s) = O(u3 ).
(8.192)
Proof. Let us drop the reference to s in h(s). We have
d iuOp(h)
e
Op e−iuh
du
Now
ieiuOp(h) Op(h)Op e−iuh − Op he−iuh .
=
(8.193)
i
Op(h)Op(e−iuh ) = Op he−iuh + Op {h, e−iuh } + O(u2 ).
2
(8.194)
The second termon the right of (8.194) is zero. Therefore, (8.193) is O(u2 ). Clearly,
eiuOp(h) Op e−iuh u=0 = 1l. Integrating O(u2 ) from 0 we obtain O(u3 ). 2
Thus we can use F (s, u) := Op e−iuh(s) in (8.163), so that
U (t)
=
lim
n→∞
n
Y
t
−i n
h
Op e
jt
n
j=1
Thus
Z
U (t, x, y)
=
lim
n→∞
Z Y
n
exp
− it
nh
jt xj +xj−1
n ,
2
, pj + i(xj − xj−1 )pj
j=1
n
Y
dpj (2π)d y = x0 ,
j=1
j=1
x = xn .

Z
Z
n X
(xj −xj−1 )pj
it

· · · exp n
lim
−h
t
×
=
n−1
Y
···
dxj
n→∞
n−1
Y
n
j=0
n
Y
dpj .
×
dxj
(2π)d y = x0 ,
j=1
j=1
x = xn .
(8.195)

jt xj +xj−1
n ,
2
, pj

(8.196)
Heuristically, this is written as follows:
Z t
Z
U (t, x, y) =
Dx,y (x(·)) D (p(·)) exp i
(ẋ(s)p(s) − h(s, x(s), p(s)) ds ,
0
where [0, t] 3 s 7→ (x(s), p(s)) is an arbitrary phase space trajectory with x(0) = y, x(t) = x
and the “measure on the phase space paths” is
n−1
n dp (j − 1 ) t
Y jt Y
2 n
Dx,y (x(·)) = lim
dx
, D p(·) =
.
n→∞
n
(2π)d
j=1
j=1
57
8.9
Hamiltonians quadratic in momenta I
Assume in addition that
h(t, x, p) =
1
(p − A(t, x))2 + V (t, x).
2
(8.197)
Then
Op h(t)
1
(pi − Ai (t, x))2 + V (t, x).
2
=
Introduce
v = p − A(t, x).
The Lagrangian for (8.199) is
L(t, x, v) =
1 2
v + vA(t, x) − V (t, x).
2
Consider the phase space path integral (8.196). The exponent depends quadratically on
p. Therefore, we can integrate it out, obtaining a configuration space path integral. More
precisely, first we make the change of variables
xj +xj−1
vj = pj − A jt
,
n ,
2
and then we do the integration wrt v:
Z
U (t, x, y)
=
Z
···
lim
n→∞
exp
it
n
j −xj−1 )
t
n
j=1
1
− vj2 − V
2
jt xj +xj−1
n ,
2
Z
lim
n→∞
! n−1
Y
dxj
j=1

=
n X
(x
Z
···
exp  itn
n
X
j=1
d
×(2π itn )−n 2
L
xj +xj−1
vj + A( jt
)
n ,
2
n
Y
dvj (2π)d y = x0 ,
j=1
x = xn

jt xj + xj−1 (xj − xj−1 ) 
,
,
t
n
2
n
n−1
Y
dxj y = x , .
0
j=1
x = xn
(8.198)
Heuristically, this is written as
Z
U (t, x, y)
=
Z t
Dx,y (x(·)) exp i
L(s, x(s), ẋ(s))ds ,
0
where [0, t] 3 s 7→ x(s) is a configuration space trajectory with x(0) = y, x(t) = x and the
formal “measure on the configuration space paths” is the same as in (8.167)
58
8.10
Hamiltonians quadratic in momenta II
Suppose, more generally, that
h(t, x, p) =
1
(pi − Ai (t, x))g ij (t, x)(pj − Aj (t, x)) + V (t, x).
2
(8.199)
Then
Op h(t)
=
1
(pi − Ai (t, x))g ij (t, x)(pj − Aj (t, x)) + V (t, x)
2
1X
−
∂ i ∂ j g ij (t, x).
4 ij x x
(For brevity, [g ij ] will be denoted g −1 and [gij ] is denoted g)
Introduce
v = g −1 (t, x) p − A(t, x)
The Lagrangian for (8.199) is
L(t, x, v) =
1 i
v gij (t, x)v j + v j Aj (t, x) − V (t, x).
2
Consider the phase space path integral (8.196). The exponent depends quadratically on
p. Therefore, we can integrate it out, obtaining a configuration space path integral. More
precisely, first we do the integration wrt p(·):
Z
U (t, x, y)
=
Dx,y x(·) D p(·) exp i
Z t
ẋ(s)p(s)
0
=
=
8.11
!
−1
1
− p(s) − A(s, x(s) g
s, x(s) p(s) − A s, x(s) − V s, x(s) ds
2
Z
Z t
1
ẋ(s)g s, x(s) ẋ(s) + ẋ(s)A(s, x(s) − V s, x(s) ds
Dx,y x(·) exp i
2
0
!
Z t
1
+
Trg s, x(s) ds
2 0
Z t Z
1
(8.200)
Dx,y (x(·)) exp
iL(s, x(s), ẋ(s)) + Trg s, x(s) ds .
2
0
Semiclassical path integration
Let us repeat the most important formulas in the presence of a Planck constant ~.
Z
i t
U (t) := Texp −
H(s)ds .
(8.201)
~ 0
59
Z
U (t, x, y)
=
−1
Dx,y (x(·)) D ~
Z t
i
(ẋ(s)p(s) − h(s, x(s), p(s)) ds ,
p(·) exp
~ 0
We assume in addition that the Hamiltonian has the form (8.199), and we set
√
x(s) = xcl (s) + ~z(s),
where xcl is the classical solution such that xcl (0) = y and xcl (t) = x.
Z t
Z
1
i
−
−d
Dx,y ~ 2 x(·) exp
U (t, x, y) = ~ 2
L(s, x(s), ẋ(s))ds
~ 0
Z t
d
i
L(s, xcl (s), ẋcl (s))ds
= ~− 2 exp
~ 0
Z
Z
i t 2
× D0,0 (z(·)) exp
∂x(s) L s, xcl (s), xcl (s) z(s)z(s)
2 0
+2∂x(s) ∂ẋ(s) L s, xcl (s), xcl (s) z(s)ż(s)
!
√ 2
+∂ẋ(s) L s, xcl (s), xcl (s) ż(s)ż(s) + O( ~) ds
=~
8.12
−d
2
" Rt
#!− 12
Rt
2
∂
L
s,
x
(s),
x
(s)
∂
∂
L
s,
x
(s),
x
(s)
1
cl
cl
cl
cl
x(s)
ẋ(s)
0R t 2
R t 0 x(s)
det
2π 0 ∂x(s) ∂ẋ(s) L s, xcl (s), xcl (s)
∂
L s, xcl (s), xcl (s)
0 ẋ(s)
Z t
√ i
× exp
L(s, xcl (s), ẋcl (s))ds 1 + O( ~) .
~ 0
General Hamiltonians – Wick quantization
Let [0, t] 3 s 7→ h(s, a∗ , a) ∈ R be a time dependent classical Hamiltonian expressed in terms
of the complex coordinates. Set
∗
H(t) := Opa ,a h(t)
and U (t) as in (8.201). (We drop the tilde from h̃ and ũ, as compared with the notation of
(4.83).)
Following Lemma 8.3 we prove that
a∗ ,a
∗
h(s)
e−iuOp
− Opa ,a e−iuh(s) = O(u2 ).
(8.202)
∗
Thus we can use F (s, u) := Opa
Opa
∗
,a
,a
e−iuh(s) in (8.163), so that
u(t) := U (t)
=
lim
n→∞
60
n
Y
j=1
∗
Opa
,a
t
e−i n h
jt
n
.
Thus, by (4.92),
u(t, a∗ , a)
=
lim exp
n→∞
X
∂ak ∂a∗j
n
Y
exp − itn h
jt
n
, a∗j , aj
j=1
k>j
a = an = · · · = a1 .
Heuristically, this can be rewritten as
Z t
Z t
∗
ds+
ds− θ(s+ − s− )∂a∗ (s+ ) ∂a(s− )
u(t, a , a) = exp
0
0
Z t
∗
.
h (s, a (s), a(s)) ds × exp −i
.
(8.203)
a=a(s),t>s>0
0
Alternatively, we can use the integral formula (4.93), and rewrite (8.203) as
∗
Z
Z
···
u(t, a , a) =
exp
n−1
X
(b∗j − a∗ )(bj+1 − bj )
(8.204)
j=1
×
n
Y
exp − itn h
j=1
jt
n
, b∗j , bj
Y dbj+1 db∗j
n−1
.
(2πi)d
j=1
Heuristically, it can be rewritten as
u(t, a∗ , a)
R
R
t
D b(·) exp 0 b∗ (s) − a∗ ∂s b(s) − ih (s, b∗ (s), b(s)) ds
=
.
Rt
R
D b(·) exp 0 b∗ (s)∂s b(s)ds
(8.205)
Here, D b(·) is some kind of a “measure” on the complex trajectories.
Let us describe another derivation of (8.205), which starts from (8.203). Consider the
operator θ on L2 ([0, t]) with the integral kernel θ(s+ − s− ) Note that
∂s+ θ(s+ − s− ) = δ(s+ − s− ).
Besides, θf (0) = 0. Therefore, ∂s θ = 1l. Thus θ is the inverse (“the Green’s function”) of
the operator ∂s with the boundary condition f (0) = 0. It is an unbounded operator with
empty resolvent. It is not antiselfadjoint – its adjoint is ∂x with the boundary condition
f (t) = 0. The corresponding sesquilinear form can be written as
Z
t
a∗ (s)∂s a(s)ds.
0
Using (2.23), (8.203) can be rewritten formally as (8.205).
61
8.13
Vacuum expectation value
In particular, we have the following expression for the vacuum expectation value:
Ω|U (t)Ω
R
R
t
D a(·) exp 0 (a∗ (s)∂s a(s) − ih (s, a∗ (s), a(s))) ds
=
.
Rt
R
D a(·) exp 0 a∗ (s)∂s a(s)ds
(8.206)
For f, g ∈ Cd we will write
a∗ (f ) = ai fi ,
a(g) = ai g i .
One often tries to express everything in terms of vacuum expectation values. To this end
introduce functions
[0, t] 3 s 7→ F (s), G(s) ∈ Cd ,
and a (typically, nonphysical) Hamiltonian
H(s) + a∗ F (s) + a G(s) .
The vacuum expectation value for this Hamiltonian is called the generating function:
! !
Z t
∗
Z(F, G) = Ω|Texp − i
H(s) + a F (s) + a G(s) ds Ω .
0
Note that we can retrieve full information about U (t) from Z(F, G) by differentiation.
Indeed let
Fi (s) = fi δ(s − t), Gi (s) = gi δ(s), fi , gi ∈ Cd .
Then
m
F1 · · · Fn G1 · · · Gm ∂Fn ∂G
Z(F, G)
F =0,G=0
n−m
∗
∗
∗
= i
a (f1 ) · · · a (fn )Ω|U (t)a (g1 ) · · · a∗ (gm )Ω
To see this, assume for simplicity that
F1 (s) = · · · = Fn (s) = f δ(s − t),
and approximate the delta function:
(
1/ 0 < s < ;
δ(s) ≈
,
0
< s < t;
G1 (s) = · · · = Gm (s) = gδ(s),
(
δ(s − t) ≈
0
1/
0 < s < t − ;
.
t − < s < t;
Using these approximations, we can write
∗
Z(sF, uG) = lim Ω|e−is a(f ) U (t)e−iu a (g) Ω
→0
∗
∗
=
eisa (f ) Ω|U (t)e−iua (g) Ω .
62
Now
m
F1 · · · F1 G1 · · · G1 ∂Fn ∂G
Z(F, G)
F =0,G=0
∗
n m
isa (f )
−iua∗ (g)
= ∂s ∂u e
Ω|U (t)e
Ω s=0, u=0
∗
n
∗
m
n−m
a (f1 ) Ω|U (t)a (g1 ) Ω .
= i
8.14
Scattering operator for Wick quantized Hamiltonians
Assume now that the Hamiltonian is defined for all times and is split into a time-independent
quadratic part and a perturbation:
h(t, a∗ , a) = a∗ εa + λq(t, a∗ , a).
Set
H0
=
Opa
∗
,a
(a∗ εa) = â∗ εâ =
Q(t)
=
Opa
∗
,a
(q(t)),
X
â∗i εi âi
i
so that H(t) = H0 + λQ(t). The scattering operator is
Z ∞
S = Texp − i
HInt (t)dt ,
−∞
where the interaction Hamiltonian is
HInt (t)
∗
Setting S = Opa
=
λeitH0 Q(t)e−itH0
=
λOpa
∗
,a
q(t, eitε a∗ , e−itε a) .
,a
(s), we can write
Z ∞
Z ∞
∗
s(a , a) = exp
dt+
dt− θ(t+ − t− )∂a(t+ ) ∂a∗ (t− )
−∞
−∞
Z ∞
iεt ∗
−iεt
× exp −iλ
q t, e a (t), e
a(t) dt a∗ = a∗ (t),
−∞
a = a(t), t ∈ R
Z ∞
Z ∞
iε(t+ −t− )
= exp
dt+
dt− e
θ(t+ − t− )∂a(t+ ) ∂a∗ (t− )
−∞
−∞
Z ∞
∗
× exp −iλ
q (t, a (t), a(t)) dt eitε a∗ = a∗ (t),
−∞
R
=
itε
(8.207)
e a = a(t), t ∈ R
R
∞
D b(·) exp −∞ b∗ (t) − eiεt a∗ (∂t + iε) b(t) − e−iεt a − iλq (t, b∗ (t), b(t)) dt
.
R
R∞ D b(·) exp −∞ b∗ (t) − eiεt a∗ (∂t + iε) b(t) − e−iεt a
63
In the firtst step we made the substitution
a(t) = e−itε aInt (t),
a∗ (t) = eitε a∗Int (t),
subsequently dropping the subscript Int. Then the differential operator was represented
as a convolution involving Green’s function of the operator ∂t + iε that has the kernel
eiε(t+ −t− ) θ(t+ − t− ).
8.15
Friedrichs diagrams
Suppose that the perturbation is written as the linear combination of monomials:
+
∗
q(a , a) =
X
k+ ,k−
+
X
|α+ | = k + ,
|α− | = k −
−
wαk +,k
,α−
α+ !α− !
+
−
a∗α aα .
−
With every term wk ,k we associate a “vertex” – a dot with k − incoming legs (to the right)
and k + outgoing legs (to the left). A Friedrichs diagram is an ordered sequence of n vertices
and a family of lines. Each leg belongs to an internal or external lines. Internal lines join an
outgoing leg with an incoming leg that belongs to a later vertex. External incoming lines are
attached to an incoming line. Such a line is extended to the right of the diagram. External
outgoing lines are attached to outgoing legs – they are extended to the left of the diagram.
We will introduce several kinds of evaluations of a Friedrichs diagram. In this subsection
we introduce the time-dependent evaluation. Suppose B is a Friedrichs diagram with vertices
numbered by n, . . . , 1. For any sequence of times tn > · · · > t1 the corresponding evaluation
of B is computed as follows.
+ −
Each line (external and internal) has its index. For each vertex we put wk ,k , considered
+ −
as a tensor with the indices corresponding to its lines. We put wk ,k for each vertex and
eiε(tj+ −tj− ) for every internal line joining the j − th and j + th vertex. At incoming external
line we put e−iεtj− a, at each outgoing external line we put eiεtj+ a∗ . We multiply all these
factors. Thus we obtain a polynomial B(tn , . . . , t1 ; a∗ , a).
Each diagram comes with its symmetry factor. It is the product of several factorials.
If there are mj + ,j − lines joining the vertices j − and j + , it involves mj + ,j − !. If there are
mj − /mj + incoming/outgoing external lines from a vertex j, it involves also mj − !/mj + !. The
complete symmetry factor will be denoted by B!
We can interpret B(tn , . . . , t1 ; a∗ , a) as a product of operators. For each line we introduce
a Hilbert space isomorphic to the one-particle space of the system. We have n + 1 time
intervals
t > tn , . . . , tj+1 > t > tj , . . . , t1 > t.
For each of these intervals we have a collection of lines that are “open” in this interval. (This
should be obvious from the diagram). Within each of these intervals we consider the tensor
product of the spaces corresponding to each of the lines that are open in this interval.
+ −
wk ,k (t) is interpreted as an operator from a k − -particle space to a k + -particle space.
If it is on the jth place in the diagram, it will be denoted Wj (tj ). 1ljB will denote the
64
time
Figure 1: Various Friedrichs vertices
identity on the tensor product of spaces corresponding to the lines that pass the jth vertex.
At the left/right end we put symmetrizators corresponding to external outgoing/incoming
−
lines, denoted Θ+
B / ΘB . Between each two consecutive vertices j + 1 and j we put the free
dynamics for time tj+1 − tj , which, by the abuse of notation, will be denoted e−i(tj+1 −tj )H0 ,
and where H0 is the sum of ε for each line. For the final/initial interval we put eitn H0 /
e−it1 H0 . Thus the evaluation of B interpreted as an operator can be written as
B(tn , . . . , t1 )
so that
=
itn H0
Θ+
(Wn (tn ) ⊗ 1lnB ) e−i(tn −tn−1 )H0 · · ·
Be
×e−i(t2 −t1 )H0 W1 (t1 ) ⊗ 1l1B e−it1 H0 Θ−
B,
+
+
B(tn , . . . , t1 ; a∗ , a) = a∗α B α
65
,α−
−
(tn , . . . , t1 )aα .
t5
t
t4
t
3
t
2
1
time
Figure 2: A disconnected Friedrichs diagram
We have the identities
=
s(a∗ , a)
∞
X
X
n=0 all
=
exp
(−iλ)
n
Z
diag.
∞
X
B(tn , . . . , t1 ; a∗ , a)
dtn · · · dt1
B!
Z
···
(8.208)
tn >···>t1
X
n=0 con.
Z
(−iλ)n
B(tn , . . . , t1 ; a∗ , a)
dtn · · · dt1
B!
Z
···
diag.
!
log (Ω|SΩ) = s(0, 0)
Z
Z
∞
X
X
B(tn , . . . , t1 )
n
=
dtn · · · dt1 ,
(−iλ)
···
B!
n=0
t
>···>t
con. diag.
n
1
no ext. lines
=
s(a∗ , a)
(Ω|SΩ)
∞
X
X
n=0
=
exp
Z
(−iλ)n
linked diag.
∞
X
n=0
X
(8.209)
tn >···>t1
Z
B(tn , . . . , t1 ; a∗ , a)
···
dtn · · · dt1
66
B!
(8.210)
(8.211)
tn >···>t1
n
(−iλ)
con. linked
diag.
Z
Z
···
tn >···>t1
!
B(tn , . . . , t1 ; a∗ , a)
dtn · · · dt1 .
B!
(8.212)
In (8.208) we sum over all diagrams. To derive (8.208) we use (8.207):
=
s(a∗ , a)
Z
∞
X
1
m!
m=0
∞
X
∞
∞
Z
dt− eiε(t+ −t− ) θ(t+ − t− )∂a(t+ ) ∂a∗ (t− )
dt+
−∞
m
−∞
Z
Z
q (tn , a∗ (tn ), a(tn )) · · · q (t1 , a∗ (t1 ), a(t1 )) dtn · · · dt1 e−itε a∗ = a∗ (t),
n=0
tn >···>t1
eitε a = a(t), t ∈ R

m
Z
Z X
∞
∞
X
X
1 
=
(−iλ)n · · ·
eiε(tj+ −tj− ) ∂a(tj+ ) ∂a∗ (tj− ) 
m!
n=0
n≥j + >j − ≥1
tn >···>t1 m=0
×q (tn , a∗ (tn ), a(tn )) · · · q (t1 , a∗ (t1 ), a(t1 )) dtn · · · dt1 e−itj ε a∗ = a∗ (t ),
×
(−iλ)n
···
j
=
∞
X
(−iλ)n
n=0
Z
Z
···
X
tn >···>t1 all
Y
diag.
1
mj + ,j − !
eitj ε a = a(tj ), n ≥ j ≥ 1.
mj+ ,j−
eiε(tj+ −tj− ) ∂a(tj+ ) ∂a∗ (tj− )
×q (tn , a (tn ), a(tn )) · · · q (t1 , a∗ (t1 ), a(t1 )) dtn · · · dt1 e−itj ε a∗ = a∗ (t ),
j
eitj ε a = a(tj ), n ≥ j ≥ 1.
∗
In the sum above, for each pair of times tj − and tj + with j + > j − we first choose mj + ,j − =
0, 1, . . . , then sum over all such possibilities. Clearly, this can be represented by a diagram.
To determine the combinatorial factor, assume for simplicity that we have one degree of
mj
akj −m (t )
a(t )kj
, then we obtain (kj −m)!j .
freedom. Then if we have a factor kjj ! which is hit by ∂a(t)
This explains the combinatorial factor on a incoming external lines. Similarly for a outgoing
external line. For internal lines the combinatorial factor comes from mj + ,j − !.
In (8.209) we sum over all connected diagrams. This follows from (8.208) by an easy
combinatorial argument and is often called the Linked Cluster Theorem.
In (8.210) we sum over all connected diagrams without external lines. In (8.211) we sum
over all linked diagrams, that is, diagrams whose each connected component has at least one
external line. In (8.212) we sum over all connected diagrams without external lines. Clearly,
(8.210), (8.211) and (8.212) follow from (8.209).
8.16
Scattering operator for time-independent perturbations
+
−
+
−
Let us now assume that the monomials wk ,k (t) = wk ,k do not depend on time.
For diagrams with no external legs the integrals
Z
Z
· · · B(tn , . . . , t1 ; a∗ , a)dtn · · · dt1
(8.213)
tn >···>t1
are then infinite or zero. We will come back to these diagrams in the next subsection.
67
Consider a linked diagram. One can expect that then (8.213) is finite. Let us compute
its value.
For ξ ∈ R introduce heuristic expressions
Z +∞
−1
,
(8.214)
eiu(H0 −ξ) du = i H0 − (ξ + i0)
0
Z
0
eiu(H0 −ξ) du =
−1
,
−i H0 − (ξ − i0)
(8.215)
2πδ(H0 − ξ).
(8.216)
−∞
Z
eit(H0 −ξ) dt =
If B is a connected linked diagram, we introduce its evaluation for the scattering amplitude
at energy E as follows. Each line (external and internal) has its index. For each vertex we
+ −
put wk ,k , considered as a tensor with the indices corresponding to its lines. Between each
consecutive vertices we insert (ξ + i0 − εj1 − · · · − εjp )−1 , where each εj corresponds to each
line between these vertices. From the right we multiply by δ(ξ − εj1 − · · · − εjp )−1 aj1 . . . ajp ,
where all ε correspond to incoming external lines. From the left we multiply by δ(ξ −
εj1 − · · · − εjp )−1 a∗j1 . . . a∗jp , where all ε correspond to incoming external lines. We obtain a
polynomial denoted Bsc (ξ; a∗ , a).
Theorem 8.4. For every connected linked diagram B the corresponding integral (8.213)
equals
Z
−2πi Bsc (ξ; a∗ , a)dξ.
Proof. We compute the integrand using the operator interpretation of B(tn , . . . , t1 ):
B(tn , . . . , t1 )
itn H0
Θ+
(Wn ⊗ 1lnB ) e−i(tn −tn−1 )H0 · · ·
Be
×e−i(t2 −t1 )H0 W1 ⊗ 1l1B e−it1 H0 Θ−
B
Z
n
−iun (H0 −ξ)
=
δ(H0 − ξ)dξΘ+
···
B (Wn ⊗ 1lB ) e
×e−iu2 (H0 −ξ) W1 ⊗ 1l1B e−it1 (H0 −ξ) Θ−
B,
=
where we substituted
un := tn − tn−1 , . . . , u2 := t2 − t1 .
and used
Z
1l =
δ(H0 − ξ)dξ.
68
Now
Z
Z
B(tn , . . . , t1 )dtn · · · dt1
···
tn >···>t1
Z ∞
Z
=
=
∞
dun · · ·
dξ
×e
Z
0
−iu2 (H0 −ξ)
n−1
2π(−i)
Z
Z
∞
n
−iun (H0 −ξ)
− ξ)Θ+
···
B (Wn ⊗ 1lB ) e
du1
0
W1 ⊗
dt1 δ(H0
−∞
1l1B e−it1 (H0 −ξ) Θ−
B
n
dξδ(H0 − ξ)Θ+
B (Wn ⊗ 1lB ) H0 − (ξ − i0)
−1
···
−1
× H0 − (ξ − i0)
W1 ⊗ 1l1B δ(H0 − ξ)Θ−
B,
2
The sum of linked diagrams will be called the linked scattering operator. Formally, we
have
S
(Ω|SΩ)
∗
slink (a , a)
=
:=
∗
Slink := Opa ,a (slink ),
∞
X
exp
n=0
=
exp −2iπ
X
Z
n
···
(−iλ)
con. linked
diag.
∞
X
n=0
Z
B(tn , . . . , t1 )
dtn · · · dt1
B!
!
tn >···>t1
X
n
λ
con. linked
diag.
Z
!
Bsc (ξ)
dξ .
B!
Note that, at least diagramwise
e−itH0 ei2tH e−itH0
.
t→∞ (Ω|e−itH0 ei2tH e−itH0 Ω)
Slink = lim
(8.217)
We can make (8.217) more general, and possibly somewhat more satisfactory as follows.
We introduce a temporal switching function R 3 t 7→ χ(t) that decays fast ±∞ and χ(0) = 1.
We then replace the time independent perturbation Q by Q (t) := χ(t/)Q. Let us denote
the corresponding scattering operator by S . Then the linked scattering operator formally
is
S
Slink = lim
,
&0 (Ω|S Ω)
One often makes the choice
χ(t/) = e−|t|/ ,
(8.218)
which goes back to Gell-Mann–Low.
Note that Slink commutes with H0 . More precisely, each diagram commutes with H0 .
69
If ε > 0, we can make a heuristic argument why Slink is unitary. It is proportional to a
unitary as a formal power series. All diagrams with only outgoing legs have zero stationary
evaluation because of the conservation of the energy. Therefore, Slink Ω = Ω. Hence Slink is
unitary.
8.17
Energy shift
We still consider a time independent perturbation. Let B be a connected diagram with no
external lines. Its evaluation is invariant wrt translations in time:
B(tn , . . . , t1 ) = B(tn + s, . . . , t1 + s).
Therefore,
Z
Z
···
B(tn , . . . , t1 )dtn · · · dt1
tn >···>t1
Z
=
Z
Z
···
dt1
B(t1 + un , . . . , t1 + u2 , t1 )dun · · · du2
un >···>u2 >0
Z
=
Z
Z
···
dt1
B(un , . . . , u2 , 0)dun · · · du2 .
(8.219)
un >···>u2 >0
This is infinite if i nonzero. However, if we do not integrate wrt t1 , we typically obtain a
finite expression in (8.219). We will see below that it can be used to compute the energy
shift.
For a connected diagram with no external lines its evaluation for the energy shift is
defined as follows. Each line (external and internal) has its index. For each vertex we put
+ −
wk ,k , considered as a tensor with the indices corresponding to its lines. Between each
consecutive vertices we insert −(εj1 + · · · + εjp )−1 , where each εj corresponds to each line
between these vertices. We obtain a number denoted Bes . (es stands for the energy shift).
Let E denote the ground state energy of H, that is E := inf sp H. We assume that we
can use the heuristic formula for the energy shift
E
=
lim (2i)−1
t→−∞
d
log(Ω|e−itH0 ei2tH e−itH0 Ω),
dt
(8.220)
which follows from (8.165) if we note that E0 = 0. Then we can derive the following
diagrammatic expansion for the energy E:
Theorem 8.5 (Goldstone theorem).
E
=
∞
X
n=0
X
con. diag.
no ext. lines
70
λn
Bes
.
B!
Figure 3: Goldstone diagram
Proof.
Applying (8.210), we get
=
log(Ω|e−itH0 ei2tH e−itH0 Ω)
∞
X
X
(−iλ)n
n=0
con. diag.
no ext. lines
Z
Z
···
B(tn , . . . , t1 )
dtn · · · dt1 .
B!
0>tn >···>t1 >t
So
(2i)−1
=
∞
X
n=0
d
log(Ω|e−itH0 ei2tH e−itH0 Ω)
dt
Z
Z
X
i(−iλ)n
···
con. diag.
no ext. lines
B(tn , . . . , t2 , t)
dtn · · · dt2 .
B!
0>tn >···>t2 >t
Now introduce
u2 := t − t2 , . . . , un := tn−1 − tn .
71
Then u2 , . . . , un ≤ 0, t ≤ u2 + · · · + un ≤ 0, and
B(tn , . . . , t2 , t)
= Wn e−i(tn −tn−1 )H0 Wn−1 ⊗ 1ln−1
···
B
−i(t2 −t)H0
2
× W2 ⊗ 1lB e
W1
n−1
iun H0
= Wn e
Wn−1 ⊗ 1lB
···
iu2 H0
2
W1 .
× W2 ⊗ 1lB e
Then we replace t by −∞ and evaluate the integral using the heuristic relation (8.215).
9
9.1
Method of characteristics
Manifolds
Let X be a manifold and x ∈ X . Tx X , resp. T#
x X will denote the tangent, resp. cotangent
space at x. TX , resp. T# X will denote the tangent, resp. cotangent bundle over X .
Suppose that x = (xi ) are coordinates on X . Then we have a natural basis in TX denoted
∂xi and a natural basis in T# X , denoted dxi . Thus every vector field can be written as v =
v(x)∂x = v i (x)∂xi and every differential 1-form can be written as α = α(x)dx = αi (x)dxi .
We will use the following notation: ∂ˆx is the operator ∂x that acts on everything on the
right. ∂x acts only on the function immediately to the right. Thus the Leibniz rule can be
written as
∂ˆx f (x)g(x) = ∂x f (x)g(x) + f (x)∂x g(x).
(9.221)
ˆ
There are situations when we could use both kinds of notation: ∂x and ∂x , as in the last
term of (9.221). In such a case we make a choice based on esthetic reasons.
9.2
1st order differential equations
Let v(t, x)∂x be a vector field and f (t, x) a function, both time-dependent. Consider the
equation
∂t + v(t, x)∂x + f (t, x) Ψ(t, x) = 0,
Ψ(0, x)
To solve it one finds first the solution of
∂t x(t, y) =
x(0, y)
=
=
Ψ(x).
v(t, x(t, y))
y;
Let x 7→ y(t, x) be the inverse function.
Proposition 9.1. Set
Z
F (t, y) :=
t
f s, x(s, y) ds.
0
Then
Ψ(t, x) := e−F
t,y(t,x)
is the solution of (9.222).
72
Ψ y(t, x)
(9.222)
(9.223)
Proof. Set
Φ(t, y) := Ψ t, x(t, y) .
(9.224)
We have
∂t Φ(t, y)
=
=
∂t + ∂t x(t, y)∂x Ψ t, x(t, y)
∂t + v t, x(t, y) ∂x Ψ t, x(t, y) .
Hence (9.222) can be rewritten as
∂t + f t, x(t, y) Φ(t, y)
=
0,
Φ(0, y)
=
Ψ(y).
(9.225)
(9.225) is solved by
Φ(t, y) := e−F (t,y) Ψ(y).
2
Consider now a vector field v(x)∂x and a function f (x), both time-independent. Consider
the equation
v(x)∂x + f (x) Ψ(x) = 0.
(9.226)
Again, first one finds solutions of
∂t x(t) = v x(t) .
(9.227)
Then we try to find a manifold Z in X of codimension 1 that crosses each curve given by a
solution of (9.226) exactly once. If the field is everywhere nonzero, this should be possible
at least locally. Then we can define a family of solutions of (9.226) denoted x(t, z), z ∈ Z,
satisfying the boundary conditions
x(0, z) = z, z ∈ Z.
This gives a local parametrization
R × Z 3 (t, z) 7→ x(t, z) ∈ X .
Let x 7→ t(x), z(x) be the inverse function.
Proposition 9.2. Set
t
Z
f x(s, z) ds.
F (t, z) :=
0
Then
Ψ(t, x) := e
−F t(x),z(x)
Ψ z(x)
is the solution of (9.222).
Proof. Set Φ(t, z) := Ψ x(t, z) . Then
∂t Φ(t, z)
= ∂t x(t, z)∂x Ψ x(t, z)
= v x(t, z) ∂x Ψ x(t, z) .
73
(9.228)
Hence we can rewrite (9.226) together with the boundary conditions as
∂t + f x(t, z) Φ(t, z) = 0,
Φ(0, z)
=
Ψ(z).
(9.229)
(9.229) is solved by
Φ(t, z) := e−F (t,z) Ψ(z).
2
9.3
1st order differential equations with a divergence term
Fort a vector field v(x)∂x we define
divv(x) = ∂xi v i (x).
Note that divv(x) depends on the coordinates.
Consider a time dependent vector field v(t, x)∂x and the equation
(∂t + v(t, x)∂x + αdivv(t, x)) Ψ(t, x)
=
0,
Ψ(0, x)
=
Ψ(x),
(9.230)
Proposition 9.3. (9.230) is solved by
Ψ(t, x) := det ∂x y(t, x)
α
Ψ(y(t, x)).
(9.231)
Proof. We introduce Φ(t, y) as in (9.224) and rewrite (9.230) as
∂t + αdivv t, x(t, y) Φ(t, y) = 0,
Φ(0, y)
=
Ψ(y)
(9.232)
We have the following identity for the determinant of a matrix valued function t 7→ A(t):
∂t det A(t) = Tr ∂t A(t)A(t)−1 det A(t).
(9.233)
Therefore,
−α
∂t det ∂y x(t, y)
−α
= −αdiv∂t x(t, y) det ∂y x(t, y)
−α
= −αdivv t, x(t, y) det ∂y x(t, y)
.
Therefore, (9.232) is solved by
−α
Φ(t, y) := det ∂y x(t, y)
Ψ(y).
2
Consider again a time independent vector field v(x)∂x and the equation
(v(x)∂x + αdivv(x)) Ψ(x)
=
0.
(9.234)
We introduce the a hypersurface Z and solutions x(t, z), as described before Prop. 9.2.
74
Proposition 9.4. Set
w(x) := ∂z x t(x), z(x) .
Then the solution of (9.234) which on Z equals Ψ(z) is
−α
Ψ(x) := det[v(x), w(x)]
Ψ z(x) .
(9.235)
Note that if X is one-dimensional, so that we can locally identify it with R and v is a
−α
.
number, (9.235) becomes Ψ(x) = C v(x)
9.4
α-densities on a vector space
Let α > 0. We say that f : (Rd )d → R is an α-density, if
hf |av1 , . . . , avd i = | det a|α hf |v1 , . . . , vd i,
(9.236)
for any linear transformation a on Rd and v1 , . . . , vd ∈ Rd .
If X is a manifold, then by an α-density we understand a function on X 3 x 7→ Ψ(x)
where Ψ(x) is an α-density on Tx X .
Clearly, given coordinates x = (xi ) on X , using the basis ∂xi in TX , we can identify an
α-density Ψ with the function
x 7→ hΨ|∂x1 , . . . , ∂xd i(x),
(9.237)
which, by abuse of notation will be also denoted Ψ(x). If we use some other coordinates
x0 = x0i , then we obtain another function x0 7→ Ψ0 (x0 ). We have the transformation property
α
Ψ(x) = |∂x x0 | Ψ0 (x0 ).
(9.238)
A good mnemotechnic way to denote an α-density is to write Ψ(x)|dx|α . Note that 0densities are usual functions, 1-densities, or simply densities are measures. p1 -densities raised
to the pth power give a density, and so one can invariantly define their Lp -norm:
Z
Z 1 p
p
=
|Ψ(x)|p dx = kφkpp .
(9.239)
Ψ(x)|dx| Proposition 9.5. If v(x)∂x is a vector field, the operator
v(x)∂x + αdivv(x)
(9.240)
is invariantly defined on α-densities.
Proof. In fact, suppose we consider some other coordinates x0 . In the new coordinates
the vector field v(x)∂x becomes v 0 (x0 )∂x0 = (∂x x0 )v(x(x0 ))∂x0 . We will denote div0 v 0 the
divergence in the new coordinates. We need to show that if
Φ = | det ∂x x0 |α Φ0 ,
Ψ = | det ∂x x0 |α Ψ0 ,
then
v(x)∂x + αdivv(x) Φ = Ψ
75
is equivalent to
v 0 (x0 )∂x0 + αdiv0 v 0 (x0 ) Φ0 = Ψ0 .
We have
divv 0
=
=
v ∂ˆx | det ∂x x0 |α
=
αv k
∂xj ∂ˆ ∂x0i k
v
∂x0i ∂xj ∂xk
∂v j
∂xj ∂ 2 x0i k
+
v
j
∂x
∂x0i ∂xj ∂xk
∂ 2 x0i ∂xj
| det ∂x x0 |α + | det ∂x x0 |α v ∂ˆx .
∂xj ∂xk ∂x0i
Therefore,
v(x)∂ˆx + αdivv(x) | det ∂x x0 |α Φ0
= | det ∂x x0 |α v 0 (x0 )∂ˆx0 + αdiv0 v 0 (x0 ) Φ0 .
2
Note that (9.231) can be written as an α-density:
Ψ(t, x)|dx|α := | det ∂x y(t, x)|α Ψ(y(t, x))|dx|α
(9.241)
Also (9.235) is naturally an α-density.
10
10.1
Hamiltonian mechanics
Symplectic manifolds
Let Y be a manifold equipped with a 2-form ω ∈ ∧2 T# Y. We say that it is a symplectic
manifold iff ω is nondegenerate at every point and dω = 0.
Let (Y1 , ω1 ), (Y2 , ω2 ) be symplectic manifolds. A diffeomorphism ρ : Y1 → Y2 is called a
symplectic transformation if ρ# ω2 = ω1 .
In what follows (Y, ω) is a symplectic manifold. We will often treat ω as a linear map
from T Y to T # Y. Therefore, the action of ω on vector fields u, w will be written in at least
two ways
hω|u, wi = hu|ωwi = ωij ui wj .
The inverse of ω as a map T Y → T # Y will be denoted ω −1 . It can be treated as a
section of ∧2 T X . The action of ω −1 on 1-forms η, ξ can be written in at least two ways
hω −1 |η, ξi = hη|ω −1 ξi = ω ij ηi ξj .
If H is a function on Y, then we define its Hamiltonian field ω −1 dH. We will often
consider a time dependent Hamiltonian H(t, y) and the corresponding dynamic defined by
the Hamilton equations
∂t y(t) = ω −1 dy H(t, y(t)).
(10.242)
76
Proposition 10.1. Flows generated by Hamilton equations are symplectic
If F, G are functions on Y, then we define their Poisson bracket
{F, G} := hω −1 |dF, dGi.
Proposition 10.2. {·, ·} is a bilinear antisymmetric operation satisfying the Jacobi identity
{F, {G, H}} + {H, {F, G}} + {G, {H, F }} = 0
(10.243)
and the Leibnitz identity
{F, GH} = {F, G}H + G{F, H}.
(10.244)
Proposition 10.3. Let t 7→ y(t) be a trajectory of a Hamiltonian H(t, y). Let F (t, y) be an
observable. Then
d
F t, y(t) = ∂t F t, y(t) + {H, F } t, y(t) .
dt
In particular,
d
H t, y(t) = ∂t H t, y(t) .
dt
10.2
Symplectic vector space
The most obvious example of a symplectic manifold is a symplectic vector
space. As we discussed before, it has the form Rd ⊕ Rd with variables (x, p) = (xi ), (pj ) and the symplectic
form
ω = dpi ∧ dxi .
(10.245)
The Hamilton equations read
∂t x =
∂p H(t, x, p),
∂t p
−∂x H(t, x, p).
=
(10.246)
The Poisson bracket is
{F, G} = ∂xi F ∂pi G − ∂pi F ∂xi G.
(10.247)
Note that Prop 10.1 and 10.2 are easy in a symplectic vector space. To show that ω is
invariant under the Hamiltonian flow we compute
d
ω
dt
=
=
=
d
dp(t) ∧ dx(t)
dt
−d∂x H x(t), p(t) ∧ dx(t) + dp(t) ∧ d∂p H x(t), p(t)
−∂p ∂x H x(t), p(t) dp(t) ∧ dx(t) + dp(t) ∧ ∂x ∂p H x(t), p(t) dx(t) = 0
Proposition 10.4. The dimension of a symplectic manifold is always even. For any symplectic manifold Y of dimension 2d locally there exists a symplectomorphism onto an open
subset of Rd ⊕ Rd .
Now (10.4) implies Prop. 10.1. Similarly, to see Prop. 10.2 we first check the Jacobi and
Leibniz identity for (10.247).
77
10.3
The cotangent bundle
Let X be a manifold. We consider the cotangent bundle T# X . It is equipped with the
canonical projection π : T# X → X .
We can always cover X with open sets equipped with charts. A chart on U ⊂ X allows
us to identify U with an open subset of Rd through coordinates x = (xi )∈ Rd . T # U can be
identified with U × Rd , where we use the coordinates (x, p) = (xi ), (pj ) .
T# X is equipped with the tautological 1-form
X
θ=
pi dxi ,
(10.248)
i
(also called Liouville or Poincaré 1-form), which does not depend on the choice of coordinates. The corresponding symplectic form, called the canonical symplectic form is
X
ω = dθ =
dpi ∧ dxi .
(10.249)
i
Thus locally we can apply the formalism of symplectic vector spaces. In particular, the
Hamilton equations have the form (10.246) and the Poisson bracket (10.247).
10.4
Lagrangian manifolds
Let Y be a symplectic manifold. Let L be a submanifold of Y and iL : L → Y be its
embedding in Y. Then we say that L is isotropic iff i#
L ω = 0. We say that it is Lagrangian
if it is isotropic and of dimension d (which is the maximal possible dimesion for an isotropic
manifold). We say that L is coisotropic if the dimension of the null space of i#
L ω is maximal
possible, that is, 2d − dim L.
Theorem 10.5. Let E ∈ R. Let L be a Lagrangian manifold contained in the level set
H −1 (E) := {y ∈ Y : H(y) = E}.
Then ω −1 dH is tangent to L.
Proof. Let y ∈ Y and v ∈ Ty L. Then since L is contained in a level set of H, we have
0 = hdH|vi = −hω −1 dH|ωvi.
(10.250)
By maximality of Ty L as an isotropic subspace of Ty Y, we obtain that ω −1 dH ∈ Ty L. 2
Clearly, symplectic transformations map Lagrangian manifolds onto Lagrangian manifolds.
10.5
Lagrangian manifolds in a cotangent bundle
Proposition 10.6. Let U be an open subset of X and consider a function U 3 x 7→ S(x) ∈
R. Then
x, dS(x) : x ∈ U
(10.251)
is a Lagrangian submanifold of T # X .
78
Proof. Tangent space of (10.251) at the point (xi , ∂xj S(x)dxj ) is spanned by
vi = ∂xi , ∂xi ∂xj S(x)∂pj
Now
hω|vi , vk i =
X
∂xi ∂xj S(x) −
i,j
X
∂xk ∂xj S(x) = 0.
k,j
2
U 3 S(x) is called a generating function of the Lagrangian manifold (10.251). If U is
connected, it is uniquely defined up to an additive constant.
Suppose that L is a connected and simply connected Lagrangian submanifold. Fix
(x0 , p0 ) ∈ L. For any (x, p) ∈ L, let γ(x,p) be a path contained in L joining (x0 , p0 ) with
(x, p).
Z
T x, p :=
θ.
γ(x,p)
#
#
Using that di#
L θ = iL dθ = iL ω = 0 and the Stokes Theorem we see that the integral does
not depend on the path. We have
dT = i#
(10.252)
L θ.
If π is injective we will say that L is projectable on the base. Then we can use U := π(L)
L
to parametrize L:
U 3 x 7→ x, p(x) ∈ L.
We then define
S(x) := T x, p(x) .
We have
∂xi S(x)dxi = dS(x) = dT x, p(x) = pi dxi .
Hence x 7→ S(x) is the unique generating functon of L satisfying S x(z0 ) = 0.
Both x 7→ S(x) and L 3 (x, p) 7→ T (x, p) will be called generating functions of the
Lagrangian manifold L. To distinguish between them we may add that the former is viewed
as a function on the base and the latter is viewed as a function on L.
We can generalize the construction of T to more general Lagrangian
manifolds. We
cov
consider the universal covering Lcov → L with the base point at
x
,
p
0
0
. Recall that L
is defined as the set of homotopy classes of curves from (x0 , p0 to x, p ∈ L contained in
L. On Lcov we define the real function
Z
Lcov 3 [γ] 7→ T ([γ]) :=
θ.
(10.253)
γ
Exactly as above we see that (10.253) does not depend on the choice of γ and that (10.252)
is true.
79
10.6
Generating function of a symplectic transformations
Let (Yi , ωi ) be symplectic manifolds. We can than consider the symplectic manifold Y2 × Y1
with the symplectic form ω1 − ω2 . Let R be the graph of a diffeomorphism ρ, that is
R := (ρ(y), y) ∈ Y2 × Y1 .
(10.254)
Clearly, ρ is symplectic iff R is a Lagrangian manifold.
Assume that Yi = T# Xi . We can identify Y2 × Y1 with T# (X2 × X1 ).
Let T# X1 3 (x1 , ξ1 ) 7→ (x2 , ξ2 ) ∈ T# X2 be a symplectic transformation. A function
X2 × X1 3 (x2 , x1 ) 7→ S(x2 , x1 ).
(10.255)
is called a generating function of the transformation ρ if it satisfies
ξ2 = −∇x2 S(x2 , x1 ),
ξ1 = ∇x1 S(x2 , x1 ).
(10.256)
Note that if assume that the graph of ρ is projectable onto X2 × X1 , then we can find a
generating function.
10.7
The Legendre transformation
Let X = Rd be a vector space. Consider the symplectic vector space X ⊕ X # = Rd ⊕ Rd
with the generic variables (v, p). It can be viewed as a cotangent bundle in two ways – we
can treat either X or X # as the base. Correspondingly, to describe any Lagrangian manifold
L in X ⊕ X # we can try to use a generating function on X or on X # . To pass from one
description to the other one uses the Legendre transformation, which is described in this
subsection.
Let U be a convex set of X . Let
U 3 v 7→ S(v) ∈ R
(10.257)
be a strictly convex C 2 -function. By strict convexity we mean that for distinct v1 , v2 ∈ U,
v1 6= v2 , 0 < τ < 1,
τ S(v1 ) + (1 − τ )S(v2 ) > S τ v1 + (1 − τ )v2 .
Then
U 3 v 7→ p(v) := ∂v S(v) ∈ X #
(10.258)
is an injective function. Let Ũ be the image of (10.258). It is a convex set, because it is the
image of a convex set by a convex function. One can define the function
Ũ 3 p 7→ v(p) ∈ U
inverse to (10.258). The Legendre transform of S is defined as
S̃(p) := pv(p) − S v(p) .
80
Theorem 10.7. (1) ∂p S̃(p) = v(p).
−1
(2) ∂p2 S̃(p) = ∂p v(p) = ∂v2 S v(p)
. Hence S̃ is convex.
˜
(3) S̃(v)
= S(v).
Proof. (1)
∂p S̃(p) = v(p) + p∂p v(p) − ∂v S v(p) ∂p v(p) = v(p).
(2)
−1
−1
.
= ∂v2 S v(p)
∂p2 S̃(p) = ∂p v(p) = ∂v p v(p)
(3)
˜
S̃(v)
= vp(v) − p(v)v p(v) + S v p(v) = S(p).
2
Thus the same Lagrangian manifold has two descriptions:
v, dS(v) : v ∈ U
=
dS̃(p), p : p ∈ Ũ .
Examples.
(1) U = Rd , S(v) = 12 vmv,
Ũ = Rd , S̃(p) = 21 pm−1 p,
√
1 − v2 .
(2) U = {v ∈ Rd : |v|
<
1},
S(v)
=
−m
p
d
2
2
Ũ = R , S̃(p) = p + m ,
(3) U = R, S(v) = ev ,
Ũ =]0, ∞[, S̃(p) = p log p − p.
Note that we sometimes apply the Legendre transformation to non-convex functions. For
instance, in the first example m can be any nondegenerate matrix.
Proposition 10.8. Suppose that S depends on an additional parameter α. Then we have
the identity
∂α S α, v(α, p) = −∂α S̃(α, p).
(10.259)
Proof. Indeed,
∂α S̃(α, p)
= ∂α pv(α, p) − S α, v(α, p)
= p∂α v(α, p) − ∂α S(α, v(α, p) − ∂v S(α, v(α, p) ∂α v(α, p)
= −∂α S α, v(α, p) .
2
81
10.8
The extended symplectic manifold
Let Y be a symplectic manifold. We introduce the extended symplectic manifold as
T# R × Y = R × R × Y,
where its coordinates have generic names (t, τ, y). Here t has the meaning of time, τ of the
energy. For the symplectic form we choose
σ := −dτ ∧ dt + ω.
Let R × Y 3 (t, y) 7→ H(t, y) be a time dependent function on Y. Let ρt be the flow
generated by the Hamiltonian H(t), that is
ρt y(0) = y(t),
(10.260)
where y(t) solves
∂t y(t) = ω −1 dy H t, y(t) .
(10.261)
Set
G(t, τ, y) := H(t, y) − τ.
It will be convenient to introduce the projection
T# R × Y 3 (t, τ, y) 7→ κ(t, τ, y) := (t, y) ∈ R × Y,
that involves forgetting the variable τ . Note that κ restricted to
G−1 (0) := {(t, τ, y) : G(t, τ, y) = 0}
(10.262)
is a bijection onto R × Y. Its inverse will be denoted by κ−1 , so that
κ−1 (t, y) = t, H(t, y), y .
Proposition 10.9. Let L be a Lagrangian manifold in Y. The set
M := {(t, τ, y) : y ∈ ρt (L), τ = H(t, y), t ∈ R}
(10.263)
satisfies the following properties:
(1) M is a Lagrangian manifold in T# R × Y;
(2) M is contained in G−1 (0)
(3) we have
κ(M) ∩ {0}×Y = {0} × L;
(4) every point in κ(M) is connected to (10.264) by a curve contained in κ(M).
Besides, conditions (1)-(4) determine M uniquely.
82
(10.264)
Proof. Let (t0 , τ0 , y0 ) ∈ M. Let v be tangent to ρt0 (L) at y0 . Then
hdy H(t0 , y0 )|vi∂τ + v
(10.265)
is tangent to M. Vectors of the form (10.265) are symplectically orthogonal to one another,
because ρt0 (L) is Lagrangian. The curve t 7→ t, H t, y(t) , y(t) is contained in M. Hence the following vector is
tangent to M:
∂t + ∂t H(t0 , y0 )∂τ + ω −1 dy H(t0 , y0 ).
(10.266)
The symplectic form applied to (10.265) and (10.266) is
−hdy H(t0 , y0 )|vi + hω|v, ω −1 dy H(t0 , y0 )i = 0.
(10.267)
(10.265) and (10.266) span the tangent space of M. Hence (1) is true (M is Lagrangian).
(2), (3) and (4) are obvious.
Let us show the uniqueness of M satisfying (1), (2), (3) and (4). Let M be a Lagrangian
submanifold contained in G−1 (0) and (t0 , τ0 , y0 ) ∈ M. By Thm 10.5, the vector
σ −1 dG(t0 , τ0 , y0 )
= σ −1 ∂t H(t0 , y0 )dt + dy H(t0 , y0 ) − dτ
(10.268)
is tangent to M. But (10.268) coincides with (10.266). Hence
∂t + ω −1 dy H(t0 , y0 ).
(10.269)
is tangent to κ(M). This means that κ(M) is invariant for the Hamiltonian flow generated
by H(t, y). Consequently,
[
κ(M) ⊃
{t} × ρt (L).
(10.270)
t∈R
κ(M) cannot be larger than the rhs of (10.270), because then condition (4) would be violated.
2
10.9
Time-dependent Hamilton-Jacobi equations
Let R × T# X 3 (t, x, p) 7→ H(t, x, p) be a time-dependent Hamilonian on T# X . Let X ⊃
U 3 x 7→ S(x) be a given function. The time-dependent Hamilton-Jacobi equation equipped
with initial conditions reads
∂t S(t, x) − H t, x, ∂x S(t, x) = 0,
S(0, x)
=
S(x).
(10.271) can be reinterpreted in more geometric terms as follows: Set
G(t, τ, x, p) := τ − H(t, x, p).
83
(10.271)
Consider a Lagrangian manifold L in Y. We want to find a Lagrangian manifold M in
G−1 (0) := {(t, τ, x, p) ∈ T# R × T# X : τ − H(t, x, p) = 0}
(10.272)
such that
κ(M) ∩ {0}×T# X = {0}×L.
Here, as in the previous subsection,
κ(t, τ, x, p) := (t, x, p).
We will also use its inverse
κ−1 (t, x, p) := t, H(t, x, p), x, p .
The relationship between the two formulations is as follows. Assume that L is a generating function of L. Then the function (t, x) 7→ S(t, x) that appears in (10.271) is the
generating function of M, which for t = 0 coincides with x 7→ S(x).
Note that the geometic formulation is superior to the traditional one, because it does
not have a problem with caustics.
The Hamilton-Jacobi equations can be solved as follows. Let R 3 t 7→ x(t, y), p(t, y) ∈
T# X be the solution of the Hamilton equation with the initial conditions on the Lagrangian
manifold L:
x(0, y), p(0, y) = y, ∂y S(y) .
Then
n
o
M = κ−1 t, x(t, y), p(t, y) : (t, y) ∈ R × U .
Let us find the generating function of M. We will use s as an alternate name for the time
variable. The tautological 1-form of T# R × T# X is
−τ ds + pdx.
Fix a point y0 ∈ U. Then the generating function of M satisfying
T κ−1 0, y0 , p(0, y0 ) = S(y0 )
is given by
−1
T κ
t, x(t, y), p(t, y) = S(y0 ) +
Z
(pdx − τ ds),
γ
where γ is a curve in M joining
κ−1 0, y0 , p(0, y0 )
with
−1
κ
(10.273)
t, x(t, y), p(t, y) .
(10.274)
We can take γ as the union of two disjoint segments: γ = γ1 ∪ γ2 . γ1 is a curve in (10.272)
with the time variable equal to zero ending at
κ−1 0, y, p(0, y) .
(10.275)
84
Clearly, since ds = 0 along γ1 , we have
Z
Z
S(y0 ) +
(pdx − τ ds) = S(y0 ) +
γ1
pdx = S(y).
(10.276)
γ1
γ2 starts at (10.275), ends at (10.274), and is given by the Hamiltonian flow. More precisely,
γ2 is
[0, t] 3 s 7→ κ−1 s, x(s, y), p(s, y) .
We have
Z
(pdx − τ ds) =
Z t
p(s, y)∂s x(s, y) − H s, x(s, y), p(s, y)
ds.
(10.277)
0
γ2
Putting together (10.276) and (10.277) we obtain the formula for the generating function of
M viewed as a function on M:
T (t, y)
=
S(y)
Z t
+
p(s, y)∂s x(s, y) − H s, x(s, y), p(s, y) ds.
(10.278)
0
If we can invert y 7→ x(t, y) and obtain the function x 7→ y(t, x), then we have a generating
of M viewed as a function on the base:
S(t, x) = T t, y(t, x) .
(10.279)
10.10
The Lagrangian formalism
Given a time-dependent Hamiltonian H(t, x, p) set
v := ∂p H(t, x, p).
Suppose that we can express p in terms of t, x, v. We define then the Lagrangian
L(t, x, v) := p(t, x, v)v − H t, x, p(t, x, v)
naturally defined on TX . Thus we perform the Legendre transformation wrt p, keeping t, x
as parameters. Note that p = ∂v L(t, x, v) and ∂x H(t, x, p) = −∂x L(t, x, v). The Hamilton
equations are equivalent to the Euler-Lagrange equations:
d
x(t) = v(t),
dt
d
∂v L t, x(t), v(t) = ∂x L x(t), v(t) .
dt
Using the Lagrangian, the generating function (10.282) can be rewritten as
Z t
T (t, y) = S(y) +
L s, x(s, y), ẋ(s, y) ds.
0
85
(10.280)
(10.281)
Lagrangians often have quadratic dependence on velocities:
L(x, v) =
1 −1
vg (x)v + vA(x) − V (x).
2
(10.282)
The momentum and the velocity are related as
p = g −1 (x)v + A(x),
v = g(x) p − A(x) .
(10.283)
The corresponding Hamiltonian depends quadratically on the momenta:
H(x, p) =
10.11
1
p − A(x) g(x) p − A(x) + V (x).
2
(10.284)
Action integral
In this subsection, which is independent of Subsect. 10.9, we will rederive the formula for
the generating function of the Hamiltonian flow constructed (10.282). Unlike in Subsect.
10.9, we will use the Lagrangian formalism.
Let [0, t] 3 s 7→ x(s, α), v(s, α) ∈ TX be a family of trajectories, parametrized by an
auxiliary variable α. We define the action along these trajectories
Z t
I(t, α) :=
L(x(s, α), v(s, α))ds.
(10.285)
0
Theorem 10.10.
∂α I(t, α) = p(x(t, α), v(t, α))∂α x(t, α) − p(x(0, α), v(0, α))∂α x(0, α).
(10.286)
Proof.
Z
∂α I(t, α)
=
t
∂x L(x(s, α), ẋ(s, α))∂α x(s, α)ds
0
=
Z t
+
∂ẋ L(x(s, α), ẋ(s, α))∂α ẋ(s, α)ds
0
Z t
d
∂x L(x(s, α), ẋ(s, α)) − ∂ẋ L(x(s, α), ẋ(s, α)) ∂α x(s, α)ds
ds
0
s=t
+p(x(s, α), v(s, α))∂α x(s, α) .
s=0
2
Theorem 10.11. Let U be an open subset in X . For y ∈ U define a family of trajectories
x(t, y), p(t, y) solving the Hamilton equation and satisfying the intial conditions
x(0, y) = y,
p(0, y) = ∂y S(y).
(10.287)
Let I(t, y) be the action along these trajectories defied as in (10.285). We suppose that we
can invert the y 7→ x(t, y) obtaining the function x 7→ y(t, x). Then
S(t, x) := I t, y(t, x) + S(y(t, x))
(10.288)
86
is the solution of (10.271), and
∂x S(t, x) = p t, y(t, x) .
(10.289)
Proof. We have
∂y I(t, y) + S(y)
= p(t, y)∂y x(t, y) − p(0, y)∂y x(0, y) + ∂y S(y)
= p(t, y)∂y x(t, y).
(10.290)
= ∂x I t, y(t, x) + S y(t, x)
(10.291)
= p(t, y)∂y x(t, y)∂x y(t, x) = p(t, y).
(10.292)
Hence,
∂x S(t, x)
Now
L x(t, y), ẋ(t, y)
=
=
∂t I(t, y) = ∂t I(t, y) + S(y)
∂t S t, x(t, y) + ∂x S t, x(t, y) ẋ(t, y).
(10.293)
(10.294)
Therefore,
∂t S t, x(t, y)
=
=
L x(t, y), ẋ(t, y) − p(t, y)ẋ(t, y)
−H x(t, y), p(t, y) .
(10.295)
(10.296)
2
10.12
Completely integrable systems
Let Y be a symplectic manifold of dimension 2d. We say that functions F1 and F2 on Y are
in involution if {F1 , F2 } = 0.
Let F1 , . . . , Fm be functions on Y and c1 , . . . , cm ∈ R. Define
−1
L := F1−1 (c1 ) ∩ · · · ∩ Fm
(cm ).
(10.297)
dF1 ∧ · · · ∧ dFm 6= 0
(10.298)
We assume that
on L. Then L is a manifold of dimension 2d − m.
Proposition 10.12. Suppose that F1 , . . . , Fm are in involution and satisfy (10.298). Then
m ≤ d and L is coisotropic. If m = d, then L is Lagrangian.
Proof. We have
hdFi |ω −1 dFj i = {Fi , Fj } = 0.
Hence ω −1 dFj is tangent to L.
hω|ω −1 dFi , ω −1 dFj i = hω −1 dFi |dFj i = −{Fi , Fj } = 0.
87
Hence the tangent space of L contains an m-dimensional subspace on which ω is zero. In
the case of a 2d − m dimensional manifold this means that L is coisotropic. 2
If H is a single function on Y, we say that it is completely integrable if we can find a
family of functions in involution F1 , . . . , Fd satisfying (10.298) on Y such that H = Fd .
Note that for completely integrable H it is easy to find Lagrangian manifolds contained
in level sets of H – one just takes the sets of the form (10.297).
11
11.1
Quantizing symplectic transformations
Linear symplectic transformations
Let ρ ∈ L(Rd ⊕ Rd ). Write ρ as a 2×2 matrix and introduce a symplectic form:
a b
0 −1l
ρ=
, ω :=
.
c d
1l 0
ρ ∈ Sp(Rd ⊕ Rd ) iff
(11.299)
ρ# ωρ = ω,
which means
a# d − c# b = 1l, c# a = a# c, d# b = b# d.
(11.300)
If
x̂0i
= aij x̂j + bij p̂j ,
p̂0i
= cij x̂j + dji p̂j ,
(11.301)
then x̂0 , p̂0 satisfy the same commutation relations as x̂, p̂. We define M pc (Rd ⊕ Rd ) to be
the set of U ∈ U (L2 (Rd )) such that there exists a matrix ρ such that
U x̂i U ∗
=
x̂0i ,
U p̂i U ∗
= p̂0i .
We will say that U implements ρ. Obviously, ρ has to be symplectic, M pc (Rd ⊕ Rd ) is a
group and the map U 7→ ρ is a homomorphism.
11.2
Metaplectic group
If χ is a quadratic polynomial on Rd ⊕ Rd , then clearly eitOp(χ) ∈ M pc (Rd ⊕ Rd ) and
implements the symplectic flow given by the Hamiltonian χ. We will denote the group
generated by such maps by M p(Rd ⊕ Rd ). Every symplectic transformation is implemented
by exactly two elements of M p.
88
11.3
Generating function of a symplectic transformation
Let ρ be as above with b invertible. We then have the factorization
a b
1l 0
0
b
1l 0
ρ=
=
,
c d
e 1l
−b#−1 0
f 1l
where
(11.302)
e = db−1 = b#−1 d# ,
f = b−1 a = a# b#−1 .
are symmetric. Define
X × X 3 (x1 , x2 ) 7→ S(x1 , x2 )
Then
a
c
b
d
:=
x1
ξ1
1
1
x1 ·f x1 − x1 ·b−1 x2 + x2 ·ex2 .
2
2
=
x2
ξ2
(11.303)
iff
∇x1 S(x1 , x2 ) = −ξ1 ,
∇x2 S(x1 , x2 ) = ξ2 .
(11.304)
The function S(x1 , x2 ) is called a generating function of the symplectic transformation ρ.
It is easy to check that the operators ±Uρ ∈ M p(X # ⊕ X ) implementing ρ have the
integral kernel equal to
i
dp
±Uρ (x1 , x2 ) = ±(2πi~)− 2 − det ∇x1 ∇x2 S e− ~ S(x1 ,x2 ) .
11.4
Harmonic oscillator
As an example, we cosider the 1-dimensional harmonic oscillator with ~ = 1. Let χ(x, ξ) :=
1 2
1 2
1 2
1 2
−tOp(χ)
equals
2 ξ + 2 x . Then Op(χ) = 2 D + 2 x . The Weyl-Wigner symbol of e
w(t, x, ξ) = (ch 2t )−1 exp(−(x2 + ξ 2 )th 2t ).
(11.305)
Its integral kernel is given by
1
1
W (t, x, y) = π − 2 (sht)− 2 exp
−(x2 + y 2 )cht + 2xy
2sht
.
e−itOp(χ) has the Weyl-Wigner symbol
w(it, x, ξ) = (cos 2t )−1 exp −i (x2 + ξ 2 )tg 2t
(11.306)
and the integral kernel
W (it, x, y) = π
− 21
t
− 12 − iπ
− iπ
4
2 [π]
| sin t|
e
e
89
exp
−(x2 + y 2 ) cos t + 2xy
2i sin t
.
Above, [c] denotes the integral part of c.
We have W (it + 2iπ, x, y) = −W (it, x, y). Note the special cases
W (0, x, y)
=
δ(x − y),
W ( iπ
2 , x, y)
=
(2π)− 2 e− 4 e−ixy ,
W (iπ, x, y)
=
e− 2 δ(x + y),
W ( i3π
2 , x, y)
=
(2π)− 2 e−
1
iπ
iπ
1
i3π
4
eixy .
1
Corollary 11.1. (1) The operator with
kernel ±(2πi)− 2 e−ixy belongs to the metaplectic
0 −1
group and implements
.
1 0
(2) The
operator
with kernel ±iδ(x + y) belongs to the metaplectic group and implements
−1 0
.
0 −1
11.5
The stationary phase method
For a quadratic form B, inert B will denote the inertia of B, that is n+ − n− , where n± is
the number of positive/negative terms of B in the diagonal form.
Theorem 11.2. Let a be smooth function on X and S a function on suppa. Let x0 be a
critical point of S, that is it satisfies
∂x S(x0 ) = 0.
(For simplicity we assume that it is the only one on suppa). Then for small ~,
Z
2
d
π
i
d
i
e ~ S(x) a(x)dx ' (2π~)− 2 ei 4 inert ∂x S(x0 ) e ~ S(x0 ) a(x0 ) + O(~− 2 +1 ).
(11.307)
Proof. The left hand side of (11.2) is approximated by
Z
2
i
i
e ~ S(x0 )+ 2~ (x−x0 )∂x S(x0 )(x−x0 ) a(x0 )dx,
(11.308)
which equals the right hand side of (11.2). 2
11.6
Semiclassical FIO’s
Suppose that X2 × X1 3 (x2 , x1 ) 7→ a(x2 , x1 ), is a function called an amplitude. Let suppa 3
(x2 , x1 ) 7→ S(x2 , x1 ) be another function, which we calle a phase. We define the Fourier
integral operator with amplitude a and phase S to be the operator from Cc∞ (X1 ) to C ∞ (X2 )
with the integral kernel
1
d
i
FIO(a, S)(x2 , x1 ) = (2π~)− 2 (∇x2 ∇x1 S(x2 , x1 )) 2 e ~ S(x2 ,x1 )
90
(11.309)
We treat FIO(a, S) as a quantization of the symplectic transformation with the generating
function S. Suppose that we can solve
∇x S(x̃, x) = p
(11.310)
FIO~ (a2 , S)∗ FIO~ (a1 , S) = Op~ (b) + O(~),
(11.311)
b(x, p) = a2 (x̃(x, p), x)a1 (x̃(x, p), x).
(11.312)
obtaining (x, p) 7→ x̃(x, p). Then
where
In particular, Fourier integral operators with amplitude 1 are asymptotically unitary.
Indeed
=
=
=
FIO~ (a2 , S)∗ FIO~ (a1 , S)(x2 , x1 )
Z
p
p
i
i
dx ∂x ∂x2 S(x, x2 ) ∂x ∂x1 S(x, x1 ) a(x2 , x)a1 (x, x1 )e− ~ S(x,x2 )+ ~ S(x,x1 )
Z
i
dxb(x2 , x, x1 )e ~ p(x2 ,x,x1 )(x2 −x1 )
Z
i
dp∂p x(x2 , p, x1 )b x2 , x(x2 , p, x1 ), x1 e ~ p(x2 −x1 ) ,
where
b(x2 , x, x1 )
p(x2 , x, x1 )
=
p
p
∂x ∂x2 S(x, x2 ) ∂x ∂x1 S(x, x1 ) a(x2 , x)a1 (x, x1 ),
Z
1
∂x S(τ x2 + (1 − τ )x1 )dτ.
=
0
11.7
Composition of FIO’s
Suppose that
X × X1 3 (x, x1 ) 7→ S1 (x, x1 ), X2 × X 3 (x2 , x) 7→ S2 (x2 , x)
(11.313)
are two functions. Given x2 , x1 , we look for x(x2 , x1 ) satisfying
∇x S2 (x2 , x(x2 , x1 )) + ∇x S1 (x(x2 , x1 ), x1 ) = 0.
(11.314)
Suppose such x(x2 , x1 ) exists and is unique. Then we define
S(x2 , x1 ) := S2 (x2 , x(x2 , x1 )) + S1 (x(x2 , x1 ), x1 )
(11.315)
Suppose S1 is a generating function of a symplectic map ρ1 : T# X1 → T# X and S2 is a
generating function of a symplectic map ρ : T# X → T# X2 . Then S is a generating function
of ρ2 ◦ ρ1 .
91
Proposition 11.3.
∇x2 ∇x1 S(x2 , x1 )
=
−∇x2 ∇x S2 (x2 , x(x2 , x1 ))
(11.316)
−1
(2)
× ∇(2)
x S2 (x2 , x(x2 , x1 )) + ∇x S1 (x(x2 , x1 ), x1 )
×∇x ∇x1 S1 (x(x2 , x1 ), x1 ).
Proof. Differentiating (11.314) we obtain
(2)
(∇x2 x)(x1 , x2 ) ∇(2)
S
(x
,
x(x
,
x
))
+
∇
S
(x(x
,
x
),
x
)
2 2
2
1
1
2
1
1
x
x
+∇x ∇x2 S2 (x2 , x(x2 , x1 ))
=
0.
(11.317)
Differentiating (11.315) we obtain
∇x1 S(x1 , x2 )
=
∇x1 S1 (x(x1 , x2 ), x1 ),
∇x2 ∇x1 S(x1 , x2 )
=
(∇x2 x)(x1 , x2 )∇x ∇x1 S1 (x(x1 , x2 ), x1 ).
(11.318)
Then we use (11.317) and (11.318). 2
In addition to two phases S1 , S2 , let
X2 × X 3 (x2 , x) 7→ a2 (x2 , x),
X × X1 3 (x, x1 ) 7→ a1 (x, x1 )
(11.319)
be two amplitudes. Then we define the composite amplitude as
a(x2 , x1 ) := a2 (x2 , x(x2 , x1 ))a1 (x(x2 , x1 ), x1 ).
(11.320)
FIO~ (a2 , S2 )FIO~ (a1 , S1 ) = FIO~ (a, S) + O(~).
(11.321)
Theorem 11.4.
12
12.1
WKB method
Lagrangian distributions
Consider a quadratic form
1
1
xSx := xi Sij xj ,
2
2
and a function on Rd
i
e 2~ xSx .
(12.322)
(12.323)
Clearly, we have the identity
i
p̂i − Sij x̂j e 2~ xSx = 0,
i = 1, . . . , d.
One can say that the phase space support of (12.323) is concentrated on
{(x, p) : pi − Sij xj = 0, i = 1, . . . , d},
92
(12.324)
which is a Lagrangian subspace of Rd ⊕ Rd .
Let us generalize (12.323). Let L be an arbitrary Lagrangian subspace of Rd ⊕ Rd . Let
an
L be the set of linear functionals on Rd ⊕ Rd such that
L=
∩
φ∈Lan
Kerφ.
Every functional in Lan has the form
φ(ξ, η) = ξj xj + η j pj .
The corresponding operator on L2 (Rd ) will be decorated by a hat:
φ̂(ξ, η) = ξij x̂j + ηij p̂j .
We say that f ∈ S 0 (Rd ) is a Lagrangian distribution associated with the subspace L iff
φ(ξ, η) ∈ Lan .
φ̂(ξ, η)f = 0,
In the generic case, the intersection of L and 0 ⊕ Rd is (0, 0). We then say that the
Lagrangian subspace is projectable onto the configuration space. Then one can find a generating function of the distribution L of the form (12.322). Lagrangian distributions associated
with L are then multiples of (12.323).
The opposite case is L = 0 ⊕ Rd . Lan is then spanned by xi , i = 1, . . . , d. The corresponding Lagrangian distributions are multiples of δ(x)
12.2
Semiclassical Fourier transform of Lagrangian distributions
Consider now the semiclassical Fourier transformation, which is an operator F~ on L2 (Rd )
given by the kernel
i
(12.325)
F~ (p, x) := e− ~ xp .
Note that for all ~, (2π~)−d/2 F~ is unitary – it will be called the unitary semiclassical
Fourier transformation. Multiplied by ±id it is an element of the metaplectic group.
Consider the Lagrangian distribution
i
e 2~ xSx ,
(12.326)
with an invertible S. Then its is easy to see that the image of (12.326) under (2π~)−d/2 F~
is
−1
i
id/2 (det S −1 )1/2 e− 2~ pS p .
More generally, we can check that the semiclassical Fourier transformation in all or only a
part of the variables preserves the set of Lagrangian distributions.
93
12.3
The time dependent WKB approximation for Hamiltonians
In this subsection we describe the WKB approximation for the time-dependent Schrödinger
equation and Hamiltonians quadratic in the momenta. For simplicity we will restrict ourselves to stationary Hamiltonians – one could generalize this subsection to time-dependent
Hamiltonians.
Consider the classical Hamiltonian
H(x, p) =
1
(p − A(x))g(x)(p − A(x)) + V (x)
2
(12.327)
with the corresponding Lagrangian
L(x, v) =
1 −1
vg (x)v + vA(x) − V (x).
2
(12.328)
We quantize the Hamiltonian in the naive way:
H~ :=
1
(−i~∂ˆ − A(x))g(x)(−i~∂ˆ − A(x)) + V (x).
2
(12.329)
We look for solutions of
i~∂t Ψ~ (t, x)
=
H~ Ψ~ (t, x).
(12.330)
Ψ~ (t, x)
e ~ S(t,x) a~ (t, x),
(12.331)
We make an ansatz
Ψ~ (0, x)
=
=
i
e
i
~ S(x)
a(x).
(12.332)
i
where a(x), S(x) are given functions. We multiply the Schrödinger equation by e− ~ S(t,x)
obtaining
i~∂ˆt − ∂t S(t, x) a~ (t, x)
(12.333)
1 −1 ˆ
=
(i ~∂x + ∂x S(t, x) − A(x))g(x)(i−1 ~∂ˆx + ∂x S(t, x) − A(x)) + V (x) a~ (t, x).
2
To make the zeroth order in ~ part of (12.333) vanish we demand that
−∂t S(t, x)
=
1
(∂x S(t, x) − A(x))g(x)(∂x S(t, x) − A(x)) + V (x).
2
(12.334)
This is the Hamilton-Jacobi equation for the Hamiltonian H. Together with the initial
conditions (12.334) can be rewritten as
−∂t S(t, x) = H x, ∂x S(x) ,
(12.335)
S(0, x)
= S(x),
94
Recall that (12.335) is solved as follows. First we need to solve the equations of motion:
ẋ(t, y) = ∂p H x(t, y), p(t, y) ,
ṗ(t, y) = −∂x H x(t, y), p(t, y) ,
x(0, y)
=
y,
p(0, y)
=
∂y S(y).
We can do it in the Lagrangian formalism. We replace the variable p by v:
v(t, x) = ∂p H x, ∂x S(t, x) .
Then
ẋ(t, y)
= v(t, y),
v̇(t, y)
= ∂x L x(t, y), v(t, y) ,
x(0, y)
= y,
v(0, y)
= ∂p H y, ∂y S(y) .
Then
S t, x(t, y) = S(y) +
Z
t
L x(s, y), v(s, y) ds
0
defines the solution of (12.335) with the initial condition (12.332), provided that we can
invert y 7→ x(t, y).
We have also the equation for the amplitude:
i~
1
ˆ
ˆ
ˆ
(12.336)
∂t + (v(t, x)∂x + ∂x v(t, x)) a~ (t, x) = ∂ˆx g(x)∂ˆx a~ (t, x).
2
2
Note that for any function b
1
1
∂ˆt + (v(t, x)∂ˆx + ∂ˆx v(t, x)) (det ∂x y(t, x)) 2 b(y(t, x)) = 0
2
(12.337)
Thus setting
i
1
Ψcl (t, x) := (det ∂x y(t, x)) 2 a y(t, x) e ~ S(t,x) .
(12.338)
We solve the Schrödinger equation modulo O(~), taking into account the initial condition:
i~∂t Ψcl (t, x)
=
H~ Ψcl (t, x) + O(~2 ),
Ψcl (0, x)
=
e ~ S(x) a(x)
i
We can improve on Ψcl by setting
1
Ψ~ (t, x) := (det ∂x y(t, x)) 2
∞
X
n=0
95
i
~n bn (t, y(t, x))e ~ S(t,x) ,
(12.339)
where
∂t bn+1
b0 (y) = a(y),
1
−1
t, y(t, x) = i~ (det ∂x y(t, x)) 2 ∂ˆx g(x)∂ˆx (det ∂x y(t, x)) 2 bn t, y(t, x) .
(The 0th order yields Ψcl (t, x)). If caustics develop after some time we can use the prescription of Subsection 12.10 to pass them.
12.4
Stationary WKB metod
The WKB method can be used to compute eigenfunctions of Hamiltonians. Let H and H~
be as in (12.327) and (12.329). We would like to solve
H~ Ψ~ = EΨ~ .
We make the ansatz
i
Ψ~ (x) := e ~ S(x) a~ (x).
i
We multiply the Schrödinger equation by e− ~ S(x) obtaining
Ea~ (x)
(12.340)
1 −1 ˆ
=
(i ~∂x + ∂x S(x) − A(x))g(x)(i−1 ~∂ˆx + ∂x S(x) − A(x)) + V (x) a~ (x).
2
To make the zeroth order in ~ part of (12.340) vanish we demand that
E
=
1
(∂x S(x) − A(x))g(x)(∂x S(x) − A(x)) + V (x),
2
which is the stationary version of theHamilton-Jacobi equation, called sometimes the eikonal
equation. Set v(x) = ∂p H x, ∂x S(x) . We have the equation for the amplitude
1
i~
v(x)∂ˆx + ∂ˆx v(x) a~ (x) = ∂ˆx g(x)∂ˆx a~ (x).
2
2
We set
a~ (x) :=
∞
X
~n an (x).
(12.341)
(12.342)
n=0
Now (12.341) can be rewritten as
1
v(x)∂ˆx + ∂ˆx v(x) a0 (x) = 0,
2
1
v(x)∂ˆx + ∂ˆx v(x) an+1 (x) = i~∂ˆx g(x)∂ˆx an (x).
2
In dimension 1 we can solve (12.343) obtaining
1
a0 (x) = |v(x)|− 2 .
96
(12.343)
This leads to an improved ansatz
− 21
Ψ~ (x) = |v(x)|
∞
X
i
~n bn (x)e ~ S(x)
n=0
We obtain the chain of equations
b0 (x) = 1,
1
1
∂t bn+1 x = i~|v(x)| 2 ∂ˆx g(x)∂ˆx |v(x)|− 2 bn x).
Thus the leading approximation is
1
i
Ψ0 (x) := |v(x)|− 2 e ~ S(x) .
(12.344)
In the case of quadratic Hamiltonians we can solve for v(x) and S(x):
q
v(x) = g(x)−1 2 E − V (x) ,
q
∂x S(x) = g(x)−1 2 E − V (x) + A(x).
12.5
Conjugating quantization with a WKB phase
Lemma 12.1. The operator B~ with the kernel
Z
i
−d
2
(x − y)p dp
(2π~)
b(x, y)p exp
~
(12.345)
equals
~ ˆ
∂x b(x, x) + b(x, x)∂ˆx + i~ (∂x b(x, y) − ∂y b(x, y)) .
2i
y=x
(12.346)
Proof. We apply Theorem 5.8. 2
Theorem 12.2. Let S, h be smooth functions. Then
i
i
e− ~ S(x) Op~ (G)e ~ S(x)
= G(x, ∂x S(x))
(12.347)
~ ˆ
+
∂x ∂p G x, ∂x S(x) + ∂p G x, ∂x S(x) ∂ˆx + O(~2 ).
2i
97
Proof. The integral kernel of the left-hand side equals
Z i
x+y −d
2
(2π~)
G
, p exp
− S(x) + S(y) + (x − y)p dp
2
~
Z Z 1
d
i
x
+
y
−2
= (2π~)
, p exp (x − y) −
∂S τ x + (1 − τ )y dτ + p dp
G
2
~
0
Z Z 1
i
d
x
+
y
= (2π~)− 2
∂S τ x + (1 − τ )y dτ exp
G
,p +
(x − y)p dp
2
~
0
Z Z 1
i
d
x
+
y
∂S τ x + (1 − τ )y dτ exp
= (2π~)− 2
G
,
(x − y)p dp
(12.348)
2
~
0
Z
Z 1
i
d
x+y
∂S(τ x + (1 − τ )y)dτ exp
+ (2π~)− 2
p∂p G
,
(x − y)p dp(12.349)
2
~
0
Z Z 1
d
+ (2π~)− 2
dσ(1 − σ)pp
0
Z 1
i
x+y
, σp +
(x − y)p dp. (12.350)
×∂p ∂p G
∂S(τ x + (1 − τ )y)dτ exp
2
~
0
We have
(12.348)
=
(12.349)
=
(12.348)
=
G(x, ∂x S(x)),
~ ˆ
∂x ∂p G(x, ∂x S(x)) + ∂p G(x, ∂x S(x))∂ˆx ,
2i
O(~2 ),
where we used Lemma 12.1 to compute the second term. 2
12.6
WKB approximation for general Hamiltonians
The WKB approximation is not restricted to quadratic Hamiltonians. Using Theorem 12.2
we easily see that the WKB method works for general Hamiltonians.
One can actually unify the time-dependent and stationary WKB method into one setup.
Consider a function H on Rd ⊕ Rd having the interpretation of the Hamiltonian. We are
interested in the two basic equations of quantum mechanics:
(1) The time-dependent Schrödinger equation:
(i~∂t − Op~ (H)) Φ~ (t, x) = 0.
(12.351)
(2) The stationary Schrödinger equation:
(Op~ (H) − E) Φ~ (x) = 0
98
(12.352)
They can be written as
Op~ (G)Φ~ (x) = 0,
(12.353)
where
(1) for (12.351), instead of the variable x actually we have t, x ∈ R × Rd , instead of p we
have τ, p ∈ R × Rd and
G(x, t, p, τ ) = τ − H(x, p).
(2) for (12.351),
G(x, p) = H(x, p) − E.
In order to solve (12.353) modulo O(~) we make an ansatz
i
Φ~ (x) = e ~ S(x) a~ (x).
i
We insert Φ~ into (12.353), we multiply by e− ~ S(x) , we set
v(x) := ∂p G(x, p),
and by (12.347) we obtain
i
e− ~ S(x) Op~ (G)Φ~
= G x, ∂x S(x) a~ (x)
~ ˆ
+
∂x v(x) + v(x)∂ˆx a~ (x)
2i
+O(~2 ).
Thus we obtain the Hamilton-Jacobi equation
G(x, ∂x S(x)) = 0
and the transport equation
1 ˆ
∂x v(x) + v(x)∂ˆx a~ (x) = O(~).
2
If we choose any solution of
1 ˆ
∂x v(x) + v(x)∂ˆx a0 = 0
2
and set
i
Φcl (x) := e ~ S(x) a0 (x)
then we obtain an approximate solution:
Op~ (G)Φcl (x) = O(~).
99
12.7
WKB functions. The naive approach
Distributions associated with a Lagrangian subspaces have a natural generalization to Lagrangian manifolds in a cotangent bundle.
Let X be a manifold and L a Lagrangian manifold in T# X . First assume that L is
projectable onto U ⊂ X and U 3 x 7→ S(x) is a generating function of L. Then
i
U 3 x 7→ a(x)e ~ S(x)
(12.354)
is a function that semiclassically is concentrated in L.
Suppose now that L is not necessarily
projectable. Then we can consider its covering
Lcov parametrized by z 7→ x(z), p(z) ∈ Lcov . Let T be a generating function of L viewed
as a function on Lcov . We would like to think of (12.354) as derived from a half-density on
the Lagrangian manifold
i
z 7→ b x(z), p(z) |dz|1/2 e ~ T (x(z),p(z)) .
(12.355)
where b is a nice function on Lcov .
If a piece of Lcov is projectable over U ⊂ X , then we can express (12.355) in terms of x:
1/2 i
(12.356)
U 3 x 7→ b x, p(z(x)) det ∂x z(x) e ~ T x,p(z(x)) |dx|1/2 .
(12.356) is actually not quite correct – there is a problem along the caustics, which should
be corrected by the so-called Maslov index.
12.8
Semiclassical Fourier transform of WKB functions
Let us apply (2π~)−d/2 F~ to a function given by the WKB ansatz:
i
Ψ~ (x) := a(x)e ~ S(x) .
Thus we consider
(2π~)−d/2
Z
i
a(x)e ~
S(x)−xp
(12.357)
dx.
We apply the stationary phase method. Given p we define x(p) by
∂x (S(x(p)) − x(p)p) = ∂x S(x(p)) − p = 0.
We assume that we can invert this function obtaining p 7→ x(p). Note that
−1
∂p x(p) = ∂x2 S x(p)
,
so locally it is possible if ∂x2 S is invertible. Let p 7→ S̃(p) denote the Legendre transform of
x 7→ S(x), that is
S̃(p) = px(p) − S x(p) .
100
Then by the stationary phase method
(2π~)−d/2 F~ Ψ~ (p)
=
e
=
e
2 S(x(p))
iπinert ∂x
4
2 S̃(p)
iπinert ∂p
4
i
|∂x2 S x(p) |−1/2 e− ~ S̃(p) a(x(p)) + O(~)
i
|∂p2 S̃(p)|1/2 e− ~ S̃(p) a(x(p)) + O(~).
One can make this formula more symmetric by replacing Ψ~ with
i
Φ~ (x) := |∂x2 S(x)|1/4 a(x)e ~ S(x) .
(12.358)
Then
(2π~)−d/2 F~ Φ~ (p)
12.9
=
e
2 S̃(p)
iπinert ∂p
4
i
|∂p2 S̃(p)|1/4 e− ~ S̃(p) a(x(p)) + O(~).
WKB functions in a neighborhood of a fold
Let us consider R × R and the Lagrangian manifold given by x = −p2 . Note that it is
not projectable in the x coordinates. It is however projectable in the p coordinates. Its
3
generating function in the p coordinates is p 7→ p3 .
We consider a function given in the p variables by the WKB ansatz
3
i p
3
1
(2π~)− 2 F~ Ψ~ (p) = e ~
Then
1
Ψ~ (x) = (2π~)− 2
Z
i
e~(
p3
3
b(p).
+xp)
b(p)dp.
(12.359)
(12.360)
√
The stationary phase method gives for x < 0, p(x) = ± −x. Thus, for x < 0,
Ψ~ (x) '
3
√
iπ
i2
1
e 4 − ~3 (−x) 2 (−x)− 4 b(− −x)
3
√
iπ
i2
1
+e− 4 + ~3 (−x) 2 (−x)− 4 b( −x).
(12.361)
iπ
Thus we see that the phase jumps by e 2 .
For x > 0 the non-stationary method gives Ψ~ (x) ' O(~∞ ). If b is analytic, we can
apply the steepest descent method to obtain
3
√
2
1
Ψ~ (x) ' e− ~3 x 2 x− 4 b(i x)
(12.362)
Note that the stationary phase and steepest descent method are poor in a close vicinity
of the fold – they give a singular behavior, even though in reality the function is continuous.
It can be approximated by replacing b(p) with b(0) in terms of the Airy function
Z ∞
i 2
1
e 3 p +ipx dp.
Ai(x) :=
2π −∞
In fact,
Ψ~ (x) ≈ b(0)(2π)1/2 ~−1/6 Ai(~−2/3 x).
101
12.10
Caustics and the Maslov correction
Let us go back to the construction described in Subsection 12.7. Recall that we had problems
with the WKB approximation near a point where the Lagrangian manifold is not projectable.
There can be various behaviors of L near such point, but it is enough to assume that we have
a simple fold. We can then represent locally the manifold as X = R × X⊥ with coordinates
(x1 , x⊥ ). The corresponding coordinates on the cotangent bundle T# X = R × R × T# X⊥
are (x1 , p1 , x⊥ , p⊥ ).
Suppose that we have a Lagrangian manifold that locally can be parametrized by (p1 , x⊥ )
with a generating function (p1 , x⊥ ) 7→ T (p1 , x⊥ ), but is not projectable on X . More precisely,
we assume that it projects to the left of x1 = 0, where it has a fold. Thus it has two sheets
given by
{x = (x1 , x⊥ ) : x1 ≤ 0} 3 x 7→ p± (x).
By applying the Legendre transformation in x1 we obtain two generating functions
{(x1 , x⊥ ) : x1 ≤ 0} 3 x 7→ S ± (x).
Suppose that we start from a function given by
i
Φ~ (p1 , x⊥ ) = e ~ T (p1 ,x⊥ )+iα b(p1 , x⊥ ),
where α is a certain phase. If we apply the unitary semiclassical Fourier transformation wrt
the variable p1 we obtain
Ψ~ (x)
=
21
−
b p−
1 (x), x⊥ det ∂x1 p1 (x)
21
π
i +
+
+ O(~).
+e ~ S (x1 ,x⊥ )+iα+i 4 b p+
1 (x), x⊥ det ∂x1 p1 (x)
i
e~S
−
(x1 ,x⊥ )+iα−i π
4
(12.363)
(12.364)
π
Thus the naive ansatz is corrected by the factor of ei 2 .
In the case of a general Lagrangian manifold, we can slightly deform it so that we can
reach each point by passing caustics only through simple folds.
12.11
Global problems of the WKB method
Let us return to the setup of Subsection 12.6. Note that the WKB method gives only a
local solution. To find a global solution we need to look for a Lagrangian manifold L in
G−1 (0). Suppose we found such a manifold. We divide it into projectable patches Li such
that π(Li ) = Ui . For each of these patches on Ui we can write the WKB ansatz
i
e ~ S(x) a(x).
Then we try to sew them together using the Maslov conditon.
This might work in the time dependent case. In fact, we can choose a WKB ansatz
corresponding to a projectable Lagrangian manifold at time t = 0, with a well defined
generating function. For small times typically the evolved Lagrangian manifold will stay
projectable and the WKB method will work well. Then caustics may form – we can then
102
consider the generating function viewed as a (univalued) function on the Lagrangian manifold
and use the Maslov prescription.
When we apply the WKB method in more than 1 dimension for the stationary Schrödinger
equation, problems are more serious. First, it is not obvious that we will find a Lagrangian
manifold. Even if we find it, it is typically not simply connected. In principle we should use
its universal covering. Thus above a single x we can have contributions from various sheets
of Lcov – typically, infinitely many of them. They may cause “destructive interference”.
12.12
Bohr–Sommerfeld conditions
The stationary WKB method works well in the special case of X = R. Typically, a
Lagrangian manifold coincides in this case with a connected component of the level set
{(x, p) ∈ R × R : H(x, p) = E}. The transport equation has a univalued solution. L is
topologically a circle, and it is the boundary of a region D, which is topologically a disc.
(This
equips
L with an orientation). The function T after going around L increases by
R
R
θ = D ω. Suppose that L crosses caustics only at simple folds, n+ of them in the “posL
itive” direction and n− in the “negative” direction. Clearly, n+ − n− = 2. (In fact, in a
typical case, such as that of a circle, we have n+ = 2, n− = 0). Then when we come back
to the initial point the WKB solution changes by
i
e~
R
D
ω−iπ
.
(12.365)
If (12.365) is different from 1, then going around we obtain contributions to WKB that
intefere destructively. Thus (12.365) has to be 1. This leads to the condition
Z
1
ω − π = 2πn, n ∈ Z,
(12.366)
~ D
or
1
2π
Z
1
ω =~ n+
2
D
,
(12.367)
which is the famous Bohr-Sommerfeld condition.
Contents
1 Introduction
1.1 Basic classical mechanics . . . .
1.2 Basic quantum mechanics . . .
1.3 Concept of quantization . . . .
1.4 The role of the Planck constant
1.5 Aspects of quantization . . . .
.
.
.
.
.
1
1
2
3
3
4
2 Preliminary concepts
2.1 Fourier tranformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
2.2 Semiclassical Fourier transformation . . . . . . . . . . . . . . . . . . . . . . .
2.3 Gaussian integrals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
5
5
5
6
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
103
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
2.4
2.5
2.6
2.7
2.8
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
. 8
. 9
. 9
. 10
. 11
. . . .
. . . .
. . . .
. . . .
. . . .
space
. . . .
. . . .
. . . .
. . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
11
11
13
14
16
17
18
19
20
20
22
4 Coherent states
4.1 General coherent states . . . . . . . . . . . . . . . . . . .
4.2 Contravariant quantization . . . . . . . . . . . . . . . . .
4.3 Covariant quantization . . . . . . . . . . . . . . . . . . . .
4.4 Connections between various quantizations . . . . . . . .
4.5 Gaussian coherent vectors . . . . . . . . . . . . . . . . . .
4.6 Creation and annihilation operator . . . . . . . . . . . . .
4.7 Quantization by an ordering prescription . . . . . . . . . .
4.8 Connection between the Wick and anti-Wick quantization
4.9 Wick symbol of a product . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
23
23
24
24
24
25
26
27
29
30
5 Pseudodifferential calculus with uniform symbol classes
5.1 Gaussian dynamics on uniform symbol classes . . . . . . .
5.2 The boundedness of quantized uniform symbols . . . . . .
5.3 Semiclassical calculus . . . . . . . . . . . . . . . . . . . . .
5.4 Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . .
5.5 Semiclassical asymptotics of the dynamics . . . . . . . . .
5.6 Frequency set . . . . . . . . . . . . . . . . . . . . . . . . .
5.7 Properties of the frequency set . . . . . . . . . . . . . . .
5.8 Algebras with a filtration/gradation . . . . . . . . . . . .
5.9 Algebra of semiclassical operators . . . . . . . . . . . . . .
5.10 Principal and subprincipal symbols . . . . . . . . . . . . .
5.11 Algebra of formal semiclassical operators . . . . . . . . . .
5.12 More about star product . . . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
31
31
31
32
33
34
35
36
36
37
37
38
38
3 x, p3.1
3.2
3.3
3.4
3.5
3.6
3.7
3.8
3.9
3.10
Gaussian integrals in complex
Integral kernel of an operator
Tempered distributions . . . .
Hilbert-Schmidt operators . .
Trace class operators . . . . .
variables
. . . . . .
. . . . . .
. . . . . .
. . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
and Weyl-Wigner quantizations
x, p-quantization . . . . . . . . . . . . . . . . . . . . . . .
Weyl-Wigner quantization . . . . . . . . . . . . . . . . . .
Star product . . . . . . . . . . . . . . . . . . . . . . . . .
Weyl operators . . . . . . . . . . . . . . . . . . . . . . . .
Weyl-Wigner quantization in terms of Weyl operators . .
Classical and quantum mechanics over a symplectic vector
Weyl quantization for a symplectic vector space . . . . . .
Positivity . . . . . . . . . . . . . . . . . . . . . . . . . . .
Parity operator . . . . . . . . . . . . . . . . . . . . . . . .
Special classes of symbols . . . . . . . . . . . . . . . . . .
104
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
6 Spectral asymptotics
39
6.1 Functional calculus and Weyl asymptotics . . . . . . . . . . . . . . . . . . . . 39
6.2 Weyl asymptotics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
6.3 Energy of many fermion systems . . . . . . . . . . . . . . . . . . . . . . . . . 42
7 Standard pseudodifferential calculus on Rd
7.1 Classes of symbols . . . . . . . . . . . . . .
7.2 Classes of pseudodifferential operators . . .
7.3 Extended principal symbol . . . . . . . . . .
7.4 Diffeomorphism invariance . . . . . . . . . .
7.5 Ellipticity . . . . . . . . . . . . . . . . . . .
7.6 Asymptotics of the dynamics . . . . . . . .
7.7 Singular support . . . . . . . . . . . . . . .
7.8 Preservation of the singular support . . . .
7.9 Wave front . . . . . . . . . . . . . . . . . .
7.10 Properties of the wave front set . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
44
44
45
46
46
46
46
47
47
47
48
8 Path integrals
8.1 Evolution . . . . . . . . . . . . . . . . . . . . . . . . . . . .
8.2 Scattering operator . . . . . . . . . . . . . . . . . . . . . . .
8.3 Bound state energy . . . . . . . . . . . . . . . . . . . . . . .
8.4 Path integrals for Schrödinger operators . . . . . . . . . . .
8.5 Example–the harmonic oscillator . . . . . . . . . . . . . . .
8.6 Path integrals for Schrödinger operators with the imaginary
8.7 Wiener measure . . . . . . . . . . . . . . . . . . . . . . . . .
8.8 General Hamiltonians – Weyl quantization . . . . . . . . . .
8.9 Hamiltonians quadratic in momenta I . . . . . . . . . . . .
8.10 Hamiltonians quadratic in momenta II . . . . . . . . . . . .
8.11 Semiclassical path integration . . . . . . . . . . . . . . . . .
8.12 General Hamiltonians – Wick quantization . . . . . . . . . .
8.13 Vacuum expectation value . . . . . . . . . . . . . . . . . . .
8.14 Scattering operator for Wick quantized Hamiltonians . . . .
8.15 Friedrichs diagrams . . . . . . . . . . . . . . . . . . . . . . .
8.16 Scattering operator for time-independent perturbations . .
8.17 Energy shift . . . . . . . . . . . . . . . . . . . . . . . . . . .
. . .
. . .
. . .
. . .
. . .
time
. . .
. . .
. . .
. . .
. . .
. . .
. . .
. . .
. . .
. . .
. . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
49
49
51
51
52
53
54
55
56
58
59
59
60
62
63
64
67
70
9 Method of characteristics
9.1 Manifolds . . . . . . . . . . . . . . . . . . . . . . . . .
9.2 1st order differential equations . . . . . . . . . . . . .
9.3 1st order differential equations with a divergence term
9.4 α-densities on a vector space . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
72
72
72
74
75
105
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
10 Hamiltonian mechanics
10.1 Symplectic manifolds . . . . . . . . . . . . . . . . . .
10.2 Symplectic vector space . . . . . . . . . . . . . . . .
10.3 The cotangent bundle . . . . . . . . . . . . . . . . .
10.4 Lagrangian manifolds . . . . . . . . . . . . . . . . .
10.5 Lagrangian manifolds in a cotangent bundle . . . . .
10.6 Generating function of a symplectic transformations
10.7 The Legendre transformation . . . . . . . . . . . . .
10.8 The extended symplectic manifold . . . . . . . . . .
10.9 Time-dependent Hamilton-Jacobi equations . . . . .
10.10The Lagrangian formalism . . . . . . . . . . . . . . .
10.11Action integral . . . . . . . . . . . . . . . . . . . . .
10.12Completely integrable systems . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
76
76
77
78
78
78
80
80
82
83
85
86
87
11 Quantizing symplectic transformations
11.1 Linear symplectic transformations . . . . . . . . .
11.2 Metaplectic group . . . . . . . . . . . . . . . . . .
11.3 Generating function of a symplectic transformation
11.4 Harmonic oscillator . . . . . . . . . . . . . . . . . .
11.5 The stationary phase method . . . . . . . . . . . .
11.6 Semiclassical FIO’s . . . . . . . . . . . . . . . . . .
11.7 Composition of FIO’s . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
88
88
88
89
89
90
90
91
12 WKB method
12.1 Lagrangian distributions . . . . . . . . . . . . . . . . . . . .
12.2 Semiclassical Fourier transform of Lagrangian distributions
12.3 The time dependent WKB approximation for Hamiltonians
12.4 Stationary WKB metod . . . . . . . . . . . . . . . . . . . .
12.5 Conjugating quantization with a WKB phase . . . . . . . .
12.6 WKB approximation for general Hamiltonians . . . . . . . .
12.7 WKB functions. The naive approach . . . . . . . . . . . . .
12.8 Semiclassical Fourier transform of WKB functions . . . . .
12.9 WKB functions in a neighborhood of a fold . . . . . . . . .
12.10Caustics and the Maslov correction . . . . . . . . . . . . . .
12.11Global problems of the WKB method . . . . . . . . . . . .
12.12Bohr–Sommerfeld conditions . . . . . . . . . . . . . . . . .
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
.
92
92
93
94
96
97
98
100
100
101
102
102
103
106
.
.
.
.
.
.
.