Ball Walk Hessian manifold Suggested algorithm

Recent Progress On Sampling Problem
Yin Tat Lee (MSR/UW), Santosh Vempala (Gatech)
My Dream
Tell the complexity of a convex problem by looking at the formula.
Example
Minimum Cost Flow Problem:
β–ͺ This is a linear program, each row has two non-zero.
It can be solved in 𝑂(π‘š 𝑛). [LS14] (Previous: 𝑂(π‘š π‘š) for graph with π‘š edges and 𝑛 vertices.)
My Dream
Tell the complexity of a convex problem by looking at the formula.
Example
Submodular Minimization:
Fundamental in combinatorial
optimization.
Worth β‰₯ 2 Fulkerson prizes
where 𝑓 satisfies diminishing return, i.e.
𝑓 𝑆+𝑒 βˆ’π‘“ 𝑆 ≀𝑓 𝑇+𝑒 βˆ’π‘“ 𝑇
β–ͺ 𝑓 can be extended to a convex function on 0,1 𝑛 .
β–ͺ subgradient of 𝑓 can be computed in 𝑛2 time.
It can be solved in 𝑂(𝑛3 ). [LSW15] (Previous: 𝑂(𝑛5 ))
βˆ€π‘‡ βŠ‚ 𝑆, 𝑒 βˆ‰ 𝑆.
Algorithmic Convex Geometry
To describe a formula, we need some operations.
Given a convex set 𝐾, we have following operations
β–ͺ Membership(x): Check if π‘₯ ∈ 𝐾.
β–ͺ Separation(x): Assert π‘₯ ∈ 𝐾, or find a hyperplane separate π‘₯ and 𝐾.
β–ͺ Width(c): Compute min𝑐 𝑇 π‘₯.
π‘₯∈𝐾
β–ͺ Optimize(c): Compute argmin𝑐 𝑇 π‘₯.
π‘₯∈𝐾
β–ͺ Sample(g): Sample according to 𝑔 π‘₯ 1𝐾 . (assume 𝑔 is logconcave)
β–ͺ Integrate(g): Compute
𝐾
𝑔 π‘₯ 𝑑π‘₯. (assume 𝑔 is logconcave)
Theorem: They are all equivalent by polynomial time algorithms.
One of the Major Source of Polynomial Time Algorithms!
Algorithmic Convex Geometry
Traditionally viewed as impractical.
Now, we have an efficient version of
ellipsoid method.
Why those operations?
For any convex 𝑓, define the dual 𝑓 βˆ— 𝑐 = argmaxπ‘₯ 𝑐 𝑇 π‘₯ βˆ’ 𝑓(π‘₯), and 𝑙𝐾 = ∞1𝐾𝑐 .
Progress: We are getting the tight polynomial equivalence between left 4.
Integration
Membership
𝑙𝐾 (π‘₯)
Width
π‘™πΎβˆ— (𝑐)
Separation
πœ•π‘™πΎ (π‘₯)
Convex Optimization
Optimization
πœ•π‘™πΎβˆ— (𝑐)
𝑔 π‘₯ 𝑑π‘₯
𝐾
Sample
~𝑒 βˆ’π‘” 1𝐾
Problem: Sampling
Input: a convex set 𝐾.
Output: sample a point from the uniform distribution on K.
Generalized Problem:
Input: a logconcave distribution 𝑓
Output: sample a point according to 𝑓.
Why? useful for optimization, integration/counting, learning, rounding.
ο‚– Best way to minimize convex function with noisy value oracle.
ο‚– Only way to compute volume of convex set.
Non-trivial application: Convex Bandit
Game: For each round 𝑑 = 1,2, β‹― , 𝑇, the player
β–ͺ Adversary selects a convex loss function ℓ𝑑
Sébastien Bubeck
Ronen Eldan
β–ͺ Chooses (possibly randomly) π‘₯𝑑 from unit ball in n dim based on past observations.
β–ͺ Receives the loss/observation ℓ𝑑 π‘₯𝑑 ∈ [0,1].
β–ͺ Nothing else about ℓ𝑑 is revealed!
Measure performance by regret:
There is a good fixed action, but
β–ͺ We only learn one point each iteration!
β–ͺ Adversary can give confusing information!
The gold of standard is getting 𝑂( 𝑇).
Namely, 𝒦 1000 𝑇 is better than 𝒦 𝑇 2/3 .
Non-trivial application: Convex Bandit
Game: For each round 𝑑 = 1,2, β‹― , 𝑇, the player
β–ͺ Adversary selects a convex loss function ℓ𝑑
Sébastien Bubeck
Ronen Eldan
β–ͺ Chooses (possibly randomly) π‘₯𝑑 from unit ball in n dim based on past observations.
β–ͺ Receives the loss/observation ℓ𝑑 π‘₯𝑑 ∈ [0,1].
β–ͺ Nothing else about ℓ𝑑 is revealed!
Measure performance by regret:
After a decade of research, we have
𝑅𝑇 = 𝑛10.5 𝑇.
(The first polynomial time and regret algorithm.)
The gold of standard is getting 𝑂( 𝑇).
Namely, 𝒦 1000 𝑇 is better than 𝒦 𝑇 2/3 .
How to Input the set
Oracle Setting:
ο‚– A membership oracle: answer YES/NO to β€œπ‘₯ ∈ 𝐾”.
ο‚– A ball π‘₯ + π‘Ÿπ΅ such that π‘₯ + π‘Ÿπ΅ βŠ‚ 𝐾 βŠ‚ π‘₯ + poly 𝑛 π‘Ÿπ΅.
Explicit Setting:
ο‚– Given explicitly, such as polytopes, spectrahedrons, …
ο‚– In this talk, we focus on polytope {𝐴π‘₯ β‰₯ 𝑏}. (m = # constraints)
Outline
ο‚– Oracle Setting:
ο‚– Introduce the ball walk
ο‚– KLS conjecture and its related conjectures
ο‚– Main Result
ο‚– Explicit Setting: (original promised talk)
ο‚– Introduce the geodesic walk
ο‚– Bound the # of iteration
ο‚– Bound the cost per iteration
Sampling Problem
Input: a convex set 𝐾 with a membership oracle
Output: sample a point from the uniform distribution on K.
Conjectured Lower Bound: 𝑛2 .
Generalized Problem: Given a logconcave distribution 𝑝, sampled π‘₯ from 𝑝.
Conjectured Optimal Algorithm: Ball Walk
At π‘₯, pick random 𝑦 from π‘₯ + 𝛿𝐡𝑛 ,
if 𝑦 is in 𝐾, go to 𝑦.
otherwise, sample again
This walk may get trapped on one side if the set is not convex.
Isoperimetric constant
For any set 𝐾, we define the isoperimetric constant πœ™πΎ by
πœ™πΎ = min
𝑆
Area(πœ•π‘†)
min(vol 𝑆 , vol 𝑆 𝑐 )
πœ™ large, hard to cut the set
Theorem
Given a random point in 𝐾, we can generate another in
𝑛
𝑂( 2 2 log(1/πœ€))
𝛿 πœ™πΎ
iterations of Ball Walk where 𝛿 is step size.
β–ͺ πœ™πΎ or 𝛿 larger, mix better.
β–ͺ 𝛿 cannot be too large, otherwise, fail probability is ~1.
πœ™ small, easy to cut the set
Isoperimetric constant of Convex Set
Note that πœ™πΎ is not affine invariant and can be arbitrary small.
L
1
πœ™πΎ = 1/𝐿.
However, you can renormalize 𝐾 such that Cov 𝐾 = 𝐼.
Definition: 𝐾 is isotropic, if it is mean 0 and Cov 𝐾 = 𝐼.
Theorem: If 𝛿 <
0.001
,
𝑛
ball walk stays inside the set with constant probability.
𝑛2
Theorem: Given a random point in isotropic 𝐾, we can generate another in 𝑂( 2 log(1/πœ€))
πœ™πΎ
To make body isotropic, we can sample the body to compute covariance.
KLS Conjecture
Kannan-Lovász-Simonovits Conjecture:
For any isotropic convex 𝐾, πœ™πΎ = Ξ©(1).
If this is true, Ball Walk takes O(𝑛2 ) iter for isotropic 𝐾
(Matched the believed information theoretical lower bound.)
To get the β€œtight” reduction from membership to sampling,
it suffices to prove KLS conjecture
KLS conjecture and its related conjectures
Slicing Conjecture:
Any unit volume convex set 𝐾 has a slice with volume Ω(1).
Thin-Shell Conjecture:
For isotropic convex 𝐾, 𝔼( π‘₯ βˆ’ 𝑛 2 ) = 𝑂(1).
Generalized Levy concentration:
For logconcave distribution 𝑝, 1-Lipschitz 𝑓 with 𝔼𝑓 = 0,
β„™ |𝑓 π‘₯ βˆ’ 𝔼𝑓| > 𝑑 = exp(βˆ’Ξ© 𝑑 ).
Essentially, it is asking if all convex sets looks like ellipsoids.
Main Result
[Lovasz-Simonovits 93] πœ™ = Ξ© 1 π‘›βˆ’1/2 .
[Klartag 2006] 𝜎 = Ξ© 1 π‘›βˆ’1/2 log1/2 𝑛.
What if we cut the body by
sphere only?
𝜎𝐾 ≝
𝑛
π‘‰π‘Žπ‘Ÿ 𝑋
2
β‰₯ πœ™πΎ
[Fleury, Guedon, Paouris 2006] 𝜎 = Ξ© 1 π‘›βˆ’1/2 log1/6 𝑛 log βˆ’2 log 𝑛.
[Klartag 2006] 𝜎 = Ξ©(1)π‘›βˆ’0.4 .
[Fleury 2010] 𝜎 = Ξ©(1)π‘›βˆ’0.375 .
[Guedon, Milman 2010] 𝜎 = Ξ©(1)π‘›βˆ’0.333 .
[Eldan 2012] πœ™ = Ξ© 1 𝜎 = Ξ©(1)π‘›βˆ’0.333 .
[Lee Vempala 2016] πœ™ = Ξ© 1 π‘›βˆ’0.25 .
In particular, we have 𝑂(𝑛2.5 ) mixing for ball walk.
Do you know better
way to bound mixing
time of ball walk?
Outline
ο‚– Oracle Setting:
ο‚– Introduce the ball walk
ο‚– KLS conjecture and its related conjectures
ο‚– Main Result
ο‚– Explicit Setting:
ο‚– Introduce the geodesic walk
ο‚– Bound the # of iteration
ο‚– Bound the cost per iteration
Problem: Sampling
Input: a polytope with π‘š constraints and 𝑛 variables.
Output: sample a point from the uniform distribution on K.
{𝐴π‘₯ β‰₯ 𝑏}
Iterations Time/Iter
of matrix
Polytopes KN09
Dikin walk
π‘šπ‘›
π‘šπ‘›1.38 Cost
inversion
LV16
Ball walk
𝑛2.5
π‘šπ‘›
LV16
Geodesic walk
π‘šπ‘›0.75
π‘šπ‘›1.38
First sub-quadratic algorithm.
How does nature mix particles?
Brownian Motion.
It works for sampling on ℝ𝑛 .
However, convex set has boundary .
Option 1) Reflect it when you hit the boundary.
However, it need tiny step for discretization.
How does the nature mixes particle?
Brownian Motion.
It works for sampling on ℝ𝑛 .
However, convex set has boundary .
Option 2) Remove the boundary by blowing up.
However, this requires explicit polytopes.
Blowing Up?
Non-Uniform Distribution on Real
After blow up
Original Polytope
Real Line
The distortion makes the hard constraint
becomes β€œsoft”.
Uniform Distribution on [0,1]
Enter Riemannian manifolds
ο‚– 𝑛-dimensional manifold M is an 𝑛-dimensional surface.
ο‚– Each point 𝑝 has a tangent space 𝑇𝑝 𝑀 of dimension 𝑛, the local linear
approximation of M at 𝑝; tangents of curves in 𝑀 lie in 𝑇𝑝 𝑀.
ο‚– The inner product in 𝑇𝑝 𝑀 depends on 𝑝: 𝑒, 𝑣
𝑝
Informally, you can think it is like assigning an unit ball for every point
Enter Riemannian manifolds
ο‚– Each point 𝑝 has a linear tangent space 𝑇𝑝 𝑀.
ο‚– The inner product in 𝑇𝑝 𝑀 depends on 𝑝: 𝑒, 𝑣
ο‚– Length of a curve 𝑐: 0,1 β†’ 𝑀 is
1
𝐿 𝑐 =
0
𝑑
𝑐 𝑑
𝑑𝑑
𝑝
𝑑𝑑
𝑐(𝑑)
ο‚– Distance 𝑑(π‘₯, 𝑦) is the infimum over all paths in M between x and y.
β€œGeneralized” Ball Walk
At x, pick random y from 𝐷π‘₯ where 𝐷π‘₯ = {𝑦: 𝑑 π‘₯, 𝑦 ≀ 1}.
Hessian manifold
Hessian manifold: a subset of ℝ𝑛 with
inner product defined by 𝑒, 𝑣 𝑝 = 𝑒𝑇 𝛻 2 πœ™ 𝑝 𝑣.
For polytope π‘Žπ‘–π‘‡ π‘₯ β‰₯ 𝑏𝑖 βˆ€π‘– , we use the log barrier function
π‘š
1
πœ™ π‘₯ =
log(
)
𝑠𝑖 π‘₯
𝑖=1
ο‚– 𝑠𝑖 π‘₯ = π‘Žπ‘–π‘‡ π‘₯ βˆ’ 𝑏𝑖 is the distance from π‘₯ to constraint 𝑖
ο‚– 𝑝 blows up when π‘₯ close to boundary
ο‚– Our walk is slower when it is close to boundary.
Suggested algorithm
At x, pick random y from 𝐷π‘₯
where 𝐷π‘₯ = {𝑦: 𝑑 π‘₯, 𝑦 ≀ 1} is induced by log barrier.
Doesn’t work!
Corresponding
Hessian Manifold
(Called Dikin Ellipsoid)
random walk on
real line
Original Polytope
Converges to the boundary since the volume of β€œboundary” is +∞.
Getting Uniform Distribution
Lemma If 𝑝 π‘₯ β†’ 𝑦 = 𝑝(𝑦 β†’ π‘₯), then stationary distribution is uniform.
To make a Markov chain 𝑝 symmetric, we use
min 𝑝 π‘₯ β†’ 𝑦 , 𝑝 𝑦 β†’ π‘₯ 𝑖𝑓 π‘₯ β‰  𝑦
𝑝 π‘₯→𝑦 =
.
β‹― 𝑖𝑓 π‘₯ = 𝑦
To implement it, we sample 𝑦 according to 𝑝(π‘₯ β†’ 𝑦)
if 𝑝 π‘₯ β†’ 𝑦 < 𝑝 𝑦 β†’ π‘₯ , go to 𝑦.
Else, we go to 𝑦 with probability 𝑝 𝑦 β†’ π‘₯ /𝑝(π‘₯ β†’ 𝑦);
Stay at x otherwise.
Dikin Walk
At x, pick random y from 𝐷π‘₯
if π‘₯ βˆ‰ 𝐷𝑦 , reject 𝑦
else, accept 𝑦 with probability
[Copied from KN09]
vol 𝐷π‘₯
min(1,
vol 𝐷𝑦
).
[KN09] proved it takes 𝑂(π‘šπ‘›) steps.
Better than the previous best 𝑂 𝑛2.5 for oracle setting.
At x, pick random y from 𝐷π‘₯
if π‘₯ βˆ‰ 𝐷𝑦 , reject 𝑦
Dikin Walk and its Limitation
else, accept 𝑦 with probability min(1,
Dikin ellipsoid is fully contained in 𝐾.
Idea: Pick next step y from a blown-up Dikin ellipsoid.
Can afford to blow up by ~ 𝑛/ log π‘š . WHP 𝑦 ∈ 𝐾.
In high dimension, volume of 𝐷π‘₯ is not that smooth. (Worst case 0,1 𝑛 )
Any larger step makes the success probability exponentially small!
0,1
𝑛
is the worst case for ball walk, hit-and-run, Dikin walk .
vol 𝐷π‘₯
vol 𝐷𝑦
).
At x, pick random y from 𝐷π‘₯
if π‘₯ βˆ‰ 𝐷𝑦 , reject 𝑦
Going back to Brownian Motion
else, accept 𝑦 with probability min(1,
vol 𝐷π‘₯
vol 𝐷𝑦
The walk is not symmetric in the β€œspace”. Tendency of going to center.
Corresponding
Hessian Manifold
Original Polytope
Taking step size to 0, Dikin walk becomes a stochastic differential equation:
𝑑π‘₯𝑑 = πœ‡ π‘₯𝑑 𝑑𝑑 + 𝜎 π‘₯𝑑 π‘‘π‘Šπ‘‘
where 𝜎 π‘₯𝑑 = πœ™ β€²β€² π‘₯𝑑 βˆ’1/2 and πœ‡(π‘₯𝑑 ) is the drift towards center.
).
What is the drift? Fokker-Planck equation
The probability distribution of the SDE
𝑑π‘₯𝑑 = πœ‡ π‘₯𝑑 𝑑𝑑 + 𝜎 π‘₯𝑑 π‘‘π‘Šπ‘‘
is given by
πœ•π‘
πœ•
1 πœ•2
2 π‘₯ 𝑝 π‘₯, 𝑑 .
π‘₯, 𝑑 = βˆ’
πœ‡ π‘₯ 𝑝 π‘₯, 𝑑 +
𝜎
𝑑𝑑
πœ•π‘₯
2 πœ•π‘₯ 2
To make the stationary distribution constant, we need
πœ•
1 πœ•2
2
βˆ’
πœ‡ π‘₯ +
𝜎
π‘₯ =0
2
πœ•π‘₯
2 πœ•π‘₯
Hence, we have πœ‡ π‘₯ = πœŽπœŽβ€².
A New Walk
A new walk:
π‘₯𝑑+β„Ž = π‘₯𝑑 + β„Ž β‹… πœ‡ π‘₯𝑑 + 𝜎 π‘₯𝑑 π‘Š
with π‘Š~𝑁(0, β„ŽπΌ).
It doesn’t make sense.
Exponential map
ο‚– Exponential map exp𝑝 : 𝑇𝑝 𝑀 β†’ 𝑀 is defined as
ο‚– exp𝑝 (𝑣) = 𝛾𝑣 (1),
ο‚– 𝛾𝑣 : unique geodesic (shortest path) from p with initial velocity 𝑣.
Geodesic Walk
A new walk:
π‘₯𝑑+β„Ž = expπ‘₯𝑑 (β„Ž/2 β‹… πœ‡ π‘₯𝑑 + 𝜎 π‘₯𝑑 π‘Š)
with π‘Š~𝑁(0, β„ŽπΌ).
Anyway to avoid
using filter?
However, this walk has discretization error.
So, we do a metropolis filter after.
Since our walk is complicated, the filter is super complicated.
Outline
ο‚– Oracle Setting:
ο‚– Introduce the ball walk
ο‚– KLS conjecture and its related conjectures
ο‚– Main Result
ο‚– Explicit Setting: (original promised talk)
ο‚– Introduce the geodesic walk
ο‚– Bound the # of iteration
ο‚– Bound the cost per iteration
Geodesic Walk
A new walk:
π‘₯𝑑+β„Ž = expπ‘₯𝑑 (β„Ž/2 β‹… πœ‡ π‘₯𝑑 + π‘Š)
with π‘Š~𝑁(0, β„ŽπΌ).
Geodesic is better than β€œstraight line”:
1) It extends infinitely.
2) It gives a massive cancellation.
Key Lemma 1: Provable Long Geodesic
Straight line defines finitely; Geodesic defines infinitely.
Thm [LV16]: For manifold induced by log barrier,
a random geodesic 𝛾 starting from π‘₯ satisfies
1
βˆ’4
π‘Žπ‘–π‘‡ 𝛾 β€² 𝑑 ≀ 𝑂(𝑛 )(π‘Žπ‘–π‘‡ π‘₯ βˆ’ 𝑏) for 0 ≀ 𝑑 ≀ 𝑂(𝑛1/4 ).
Namely, the geodesic is well behavior for a long time.
Remark:
If central path in IPM had this, we have a π‘š5/4 time algorithm for MaxFlow!
Key Lemma 2: Massive Cancellation
Consider a SDE on 1 dimensional real line (NOT manifold)
𝑑π‘₯𝑑 = πœ‡ π‘₯𝑑 𝑑𝑑 + 𝜎 π‘₯𝑑 π‘‘π‘Šπ‘‘ .
How good is the β€œEuler method”, namely π‘₯0 + β„Žπœ‡ π‘₯0 + β„ŽπœŽ π‘₯0 π‘Š?
By β€œTaylor” expansions, we have
β„Ž β€²
π‘₯β„Ž = π‘₯0 + β„Žπœ‡ π‘₯0 + β„ŽπœŽ π‘₯0 π‘Š + 𝜎 π‘₯0 𝜎 π‘₯0 π‘Š 2 βˆ’ 1 + 𝑂 β„Ž1.5 .
2
If 𝜎 β€² π‘₯0 β‰  0, the error is 𝑂(β„Ž).
If 𝜎 β€² π‘₯0 = 0, the error is 𝑂(β„Ž1.5 ).
For geodesic walk, 𝜎 β€² π‘₯0 = 0 (Christoffel symbols vanish in normal
coordinates)
Convergence Theorem
Thm [LV16]: For log barrier, the geodesic walk mixes in 𝑂 π‘šπ‘›0.75 steps.
Thm [LV16]: For log barrier on 0,1 𝑛 , it mixes in 𝑂(𝑛1/3 ) steps. 
The best bound for ball-walk, hit-and-run and Dikin walk
is 𝑂(𝑛2 ) steps for 0,1 𝑛 .
Our walk is similar to Milstein method.
Is high order method for
SDE used in MCMC?
Outline
ο‚– Oracle Setting:
ο‚– Introduce the ball walk
ο‚– KLS conjecture and its related conjectures
ο‚– Main Result
ο‚– Explicit Setting: (original promised talk)
ο‚– Introduce the geodesic walk
ο‚– Bound the # of iteration
ο‚– Bound the cost per iteration
Can we simply do Taylor expansion?
How to implement the algorithm
In high dim, it may take π‘›π‘˜ time to
compute the π‘˜ π‘‘β„Ž derivatives.
In tangent plane at x,
1. pick 𝑀 ∼ 𝑁π‘₯ (0, 𝐼), i.e. standard Gassian in β€–. β€–π‘₯
2. Compute 𝑦 =
β„Ž
expπ‘₯ πœ‡
2
π‘₯ + β„Žπ‘€
3. Accept with probability Min
𝑝 𝑦→π‘₯
1,
𝑝 π‘₯→𝑦
How to compute geodesic and rejection probability?
Need high accuracy for rejection probability  due to β€œdirectedness”.
Geodesic is given by geodesic equation; probability is given by Jacobi field.
Collocation Method for ODE
A weakly polynomial time algorithm for some ODEs
Consider the ODE 𝑦 β€² = 𝑓 𝑑, 𝑦(𝑑) with 𝑦 0 = 𝑦0 .
Given a degree 𝑑 poly π‘ž and distinct points 𝑑1 , 𝑑2 , β‹― , 𝑑𝑑 ,
let 𝑇(π‘ž) be the unique degree 𝑑 poly 𝑝 s.t.
p
𝑝′ 𝑑 = 𝑓(𝑑, π‘ž(𝑑))
on 𝑑 = 𝑑1 , 𝑑2 , β‹― , 𝑑𝑑
𝑝 0 = π‘ž(0).
Lem [LV16]: 𝑇 is well defined. If 𝑑𝑖 are Chebyshev points on [0,1], then
𝐿𝑖𝑝 𝑇 = 𝑂 𝐿𝑖𝑝 𝑓 .
Thm [LV16]: If 𝐿𝑖𝑝 𝑓 ≀ 0.001, we can find a fix point of 𝑇 efficiently.
Collocation Method for ODE
A weakly polynomial time algorithm for some ODEs
Consider the ODE 𝑦 β€² = 𝑓 𝑑, 𝑦(𝑑) with 𝑦 0 = 𝑦0 .
Thm [LV16]: Suppose that
ο‚– 𝐿𝑖𝑝 𝑓 ≀ 0.001
ο‚– There is a degree 𝑑 poly 𝑝 such that 𝑝′ βˆ’ 𝑓 β€² ≀ πœ€.
Then, we can find a 𝑦 such that 𝑦 βˆ’ 𝑦 1 = 𝑂(πœ€) in time
𝑂(𝑑log 2 π‘‘πœ€ βˆ’1 ) with 𝑂(𝑑log π‘‘πœ€ βˆ’1 ) evaluations of 𝑓.
Remark: No need to compute 𝑓′!
In general, the runtime is 𝑂(𝑛𝑑Lip𝑂(1) (𝑓)) instead.
𝒕𝒉
How can I bound the πŸπŸ•πŸŽ
derivatives?
For 1 variable function, we can estimate π‘˜ π‘‘β„Ž derivatives easily.
Idea: reduce estimating derivatives of general functions to 1 variable.
In general, we write 𝐹 ≀π‘₯ 𝑓
Dπ‘˜ 𝐹(π‘₯) ≀ 𝑓 π‘˜ 0 .
Calculus rule: 𝐹 ≀π‘₯ 𝑓 and 𝐺 ≀𝐹(π‘₯) 𝑔, then 𝐺 ∘ 𝐹 ≀π‘₯ 𝑔 ∘ (𝑓 βˆ’ 𝑓 0 ).
Implementation Theorem
Using the trick before, we show geodesic can be approximated by 𝑂(1)
degree poly.
Hence, collocation method finds in 𝑂(1) steps.
Thm [LV16]: If β„Ž ≀ π‘›βˆ’1/2 , 1 step of Geodesic walk can be implemented
in matrix multiplication time.
For hypercube, β„Ž ≀ 𝑂(1) suffices.
Questions
ο‚– We have no background on numerical ODE/SDE and RG.
So, the running time should be improvable easily.
ο‚– How to avoid the filtering step?
ο‚– Is there way to tell a walk mixed or not?
(i.e. even if we cannot prove KLS, the algorithm can still stop early.)
ο‚– Is higher order method in SDE useful in MCMC?
ο‚– Any other suggestion/heuristic for sampling on convex set?