• Lower bound on deterministic evaluation
algorithms for NOR circuits
• Yao’s Principle for proving lower bounds
2006/3/8
Randomized Algorithms, Lecture 3
1
Part 0
Deterministic algorithms for
evaluating NOR circuits
2006/3/8
Randomized Algorithms, Lecture 3
2
A three-level complete NOR
circuit
x1
x2
x3
x4
y
x5
x6
x7
x8
2006/3/8
Randomized Algorithms, Lecture 3
3
Evaluating NOR circuits
Let Fk (x1 ; x2 ; : : : ; x2k ) denote the value
of the level-k complete NOR circuit on
input x1 ; x2 ; : : : ; x2k .
The NOR circuit evaluation problem:
F Input: x1 ; x2 ; : : : ; x2k .
F Output: Fk (x1 ; x2 ; : : : ; x2k ).
2006/3/8
Randomized Algorithms, Lecture 3
4
Take a closer look at NOR
x1
x2
y = F1(x1, x2)
To determine y = F1 (x1 ; x2 ), there are
two cases;
Case 1: y=0 At least one of x1 and x2 is 1, so
we only have to read one of x1 and
x2 if we are lucky.
Case 2: y=1 Both of x1 and x2 are 0, so we
have to read both x1 and x2 to
ensure y = 1.
2006/3/8
Randomized Algorithms, Lecture 3
5
Question: can a correct
evaluation algorithm avoid
reading all inputs?
X1
X2
X3
X4
X5
X6
Y
X7
X8
2006/3/8
Randomized Algorithms, Lecture 3
6
The answer
• It’s impossible for deterministic any
algorithms to correctly evaluate Fk in
sublinear (i.e., o(2k)) time for any input
x1,x2,…,x2k.
• It’s possible for a randomized algorithm to
correctly evaluate Fk in expected sublinear
(i.e., o(2k)) time for any input x1,x2,…,x2k.
2006/3/8
Randomized Algorithms, Lecture 3
7
Theorem
For any deterministic algorithm A that always
correctly evaluates Fk for any k ¸ 1, there
exists an input instance x1 ; x2 ; : : : ; x2k such
that A has to read all 2k numbers to determine
Fk (x1 ; x2 ; : : : ; x2k ).
2006/3/8
Randomized Algorithms, Lecture 3
8
Proof strategy
• Without loss of generality, we may assume
that a deterministic algorithm works in a
depth-first manner.
• Prove by induction on k that any
deterministic depth-first algorithm has two
“trouble” instances, one for output 0 and
the other for output 1.
2006/3/8
Randomized Algorithms, Lecture 3
9
Depth-first evaluation
• Definition: does not read any leaf beneath
the second leg until the value of the first
leg is determined
• For example, RandEval is an depth-first
evaluation algorithm.
• Observation: Without loss of generality, we
can focus only depth-first evaluation
algorithms. Why?
2006/3/8
Randomized Algorithms, Lecture 3
10
The reason
• Lemma. Let D’ be a deterministic algorithm that
correctly evaluates any NOR circuit. Then, there
is a deterministic depth-first evaluation algorithm
D such that
– The outputs of D and D’ coincide on any input.
– The number of leaves by D is always no more than
that by D’.
• Proof: a simple exercise.
• So, it suffices to focus on depth-first algorithms.
2006/3/8
Randomized Algorithms, Lecture 3
11
Proof by induction on k
• CLAIM: The following statement holds for
any positive integer k and any
deterministic depth-first evaluation
algorithm A.
– There is an input Xk,1 (respectively, Xk,0) such
that A has to read all numbers in the input to
correctly compute Fk(Xk,1)=1 (respectively,
Fk(Xk,0)=0).
2006/3/8
Randomized Algorithms, Lecture 3
12
Basis: k=1.
Xk;1
0
0
1
Xk;0
0
1
0
if A reads the ¯rst leg ¯rst.
Xk;0
1
0
0
if A reads the second leg ¯rst.
2006/3/8
Randomized Algorithms, Lecture 3
13
Induction step: k k+1.
Xk+1;1
Xk;0
0
1
Xk;0
2006/3/8
0
Randomized Algorithms, Lecture 3
14
Induction step: k k+1.
Xk+1;0
First leg to be evaluated
by A.c
Xk;0
0
0
Xk;1
2006/3/8
1
Randomized Algorithms, Lecture 3
15
Induction step: k k+1.
Xk+1;0
Xk;1
1
0
Xk;0
0
First leg to be evaluated
by A.
2006/3/8
Randomized Algorithms, Lecture 3
16
Theorem
For any deterministic algorithm A that always
correctly evaluates Fk for any k ¸ 1, there
exists an input instance x1 ; x2 ; : : : ; x2k such
that A has to read all 2k numbers to determine
Fk (x1 ; x2 ; : : : ; x2k ).
2006/3/8
Randomized Algorithms, Lecture 3
17
Therefore, any deterministic
algorithm takes (2k) time to
correctly evaluate Fk in the worst
case!
2006/3/8
Randomized Algorithms, Lecture 3
18
Randomization is indeed very
helpful for this problem!
2006/3/8
Randomized Algorithms, Lecture 3
19
Comments
• Throwing coins does the magic:
– For any input, including Xk,0 and Xk,1, RandEval runs
in expected O(n0.793) time.
– By Snir [TCS’85].
• Snir’s algorithm RandEval is an optimal Las
Vegas algorithm for evaluating NOR circuits.
– That is any Las Vegas algorithm requires (n0.793)
time to evaluate NOR circuits.
– Proved by Saks and Wigderson [FOCS’86].
2006/3/8
Randomized Algorithms, Lecture 3
20
Part 1
Yao’s Principle for Proving Lower
Bounds [FOCS’77]
2006/3/8
Randomized Algorithms, Lecture 3
21
Yao’s Inequality
Let ¦ be a problem. Let D consist of all deterministic algorithms for solving ¦. Let I consist of all
possible inputs for ¦. Let Ip denote the instance of
¦ with probability distribution p. Then, the following equality holds for any probability distribution p
and any Las Vegas randomized algorithm R.
·
2006/3/8
min
expected time of A running on Ip
2D
A
max
expected time of R running on I:
2I
I
Randomized Algorithms, Lecture 3
22
Outline
• An alternative way to view a randomized
algorithm.
• A little bit of game theory
– Two person game and its Nash equilibrium
– Von Neumann’s Minimax Theorem
– Loomis Theorem
• Yao’s inequality
• 牛刀小試:
– It takes (n0.694) time for any Las Vegas algorithm to
correctly evaluate a NOR circuit.
2006/3/8
Randomized Algorithms, Lecture 3
23
An alternative viewpoint
Let I be an input for the problem ¦. Let D consist
of ALL deterministic algorithms that correctly solve
¦ on I. Then each Las Vegas algorithm for ¦ is
actually a probability distribution over D .
2006/3/8
Randomized Algorithms, Lecture 3
24
For example,
• such an instance has 8 deterministic
depth-first evaluation algorithms.
"
u
x3
x4
y
"
"
#
#
"
w
"
2006/3/8
w
"
"
v
v
"
"
x1
x2
u
"
Randomized Algorithms, Lecture 3
25
Algorithm RandEval
boolean function RandEval(x1 ; : : : ; x2t ) f
if (2t == 1) return x1 ;
throw a fair coin;
if (the head appears) f
return (RandEval(x1 ; : : : ; xt ) == 1) ?
0 : ! RandEval(xt+1 ; : : : ; x2t );
g else f
return (RandEval(xt+1 ; : : : ; x2t ) == 1) ?
0 : ! RandEval(x1 ; : : : ; xt );
g
g
2006/3/8
Randomized Algorithms, Lecture 3
26
The alternative view
• RandEval is the uniform distribution over
the eight deterministic depth-first
evaluation algorithms.
2006/3/8
Randomized Algorithms, Lecture 3
27
Strategic Interactions
• Players:
• Strategies:
• Payoffs:
Reynolds and Philip Morris
{ Advertise , Do Not Advertise }
Companies’ Profits
• Each firm earns $5 million from its customers
• Advertising costs a firm $2 million
• Advertising captures $3 million from competitor
• How to represent this game?
2006/3/8
Randomized Algorithms, Lecture 3
28
Strategic Normal Form
PLAYERS
No Ad
Reynolds
Ad
Philip Morris
No Ad
Ad
5 , 5
2 , 6
6 , 2
STRATEGIES
2006/3/8
3 , 3
PAYOFFS
Randomized Algorithms, Lecture 3
29
Nash Equilibrium
Reynolds
No Ad
Ad
Philip Morris
No Ad
Ad
5 , 5
2 , 6
6 , 2
3 , 3
• Best reply for Reynolds:
• If Philip Morris advertises:
• If Philip Morris does not advertise:
advertise
advertise
• Regardless of what you think Philip Morris will do
Advertise!
2006/3/8
Randomized Algorithms, Lecture 3
30
Nash Equilibrium
j
• When the row player
uses the i-th strategy,
the best strategy for
the column player is
strategy j.
• When the column
player uses the j-th
strategy, the best
strategy for the row
player is strategy i.
i
2006/3/8
Randomized Algorithms, Lecture 3
31
Another example:
Prisoner’s Dilemma
2006/3/8
Randomized Algorithms, Lecture 3
32
The scenario
• In the Prisoner’s Dilemma, A and B are
picked up by the police and interrogated
in separate cells without the chance to
communicate with each other.
2006/3/8
Randomized Algorithms, Lecture 3
33
Both are told:
If you both confess, you will both get four years in
prison.
If neither of you confesses, the police will be able to
pin part of the crime on you, and you’ll both get two
years.
If one of you confesses but the other doesn’t, the
confessor will make a deal with the police and will go
free while the other one goes to jail for five years.
2006/3/8
Randomized Algorithms, Lecture 3
34
Payoff Table
A does not
confess
B does not
confess
B
confesses
2, 2
5, 0
0, 5
4, 4
A confesses
2006/3/8
Randomized Algorithms, Lecture 3
35
Question:
Does each game have a unique
equilibrium?
2006/3/8
Randomized Algorithms, Lecture 3
36
Matching Pennies
No Nash equilibrium
2006/3/8
Head
Tail
Head
1, -1
-1, 1
Tail
-1, 1
1, -1
Randomized Algorithms, Lecture 3
37
Go to movie together, but action
or romance?
Action
Romance
Action
2, 1
0, 0
Romance
0, 0
1, 2
2006/3/8
Randomized Algorithms, Lecture 3
38
Focus
2-person 0-sum games –
the sum of payoffs of two players
is zero in each cell of the table
2006/3/8
Randomized Algorithms, Lecture 3
39
2-person 0-sum games
c
• It suffices to list the
payoffs for the rowplayer.
r
2006/3/8
Randomized Algorithms, Lecture 3
M (r; c)
40
An observation
max min M (r; c) · min max M (r; c):
r
c
c
r
Try to prove it yourself as a simple exercise.
LHS = the maximum payo® that
the row-player can guarantee for
himself.
2006/3/8
RHS = the minimum loss that
the column-player can guarantee
for himself.
Randomized Algorithms, Lecture 3
41
For example,
max min M (r; c) = ¡1 < 1 = min max M (r; c):
r
2006/3/8
c
c
剪刀
石頭
布
剪刀
0
-1
1
石頭
1
0
-1
布
-1
1
0
Randomized Algorithms, Lecture 3
r
42
Another example
max min M (r; c) = 0 = min max M (r; c):
r
c
r
剪刀
饅頭
布
剪刀
0
1
2
饅頭
-1
0
1
布
-2
-1
0
Saddle point
2006/3/8
c
Randomized Algorithms, Lecture 3
43
Not every 2-person 0-sum game
has a saddle point
• However, with respect to “mixed
strategies”, von Neumann showed that
each 2-person 0-sum game has a saddle
point.
2006/3/8
Randomized Algorithms, Lecture 3
44
Pure strategies versus
mixed strategies
A mixed strategy is a probability distribution over
the set of all pure strategies. Let p be a mixed strategy for the row player and q be a mixed strategy
for the column player. That is, pr is the probability
for the row player to use his r-th strategy, and qc is
the probability for the column player to use his c-th
strategy. Then the expected payo® with respect to p
and q is
XX
pT M q =
pr M (r; c)qc :
r
2006/3/8
c
Randomized Algorithms, Lecture 3
45
von Neumann’s
Minimax Theorem
For any 2-person 0-sum game, we have
max min pT M q = min max pT M q:
p
q
q
p
That is, each 2-person zero-sum game has a saddle
point with respect to mixed strategies.
2006/3/8
Randomized Algorithms, Lecture 3
46
Loomis’ Theorem
For any 2-person 0-sum game, we have
max min pT M ec = min max eTr M q;
p
c
q
r
where ei means running the i-th strategy with probability
1.
To see the theorem, just observe that when p is known,
the column player has an optimal strategy that is a pure
strategy. A similar observation holds for the row player,
too.
2006/3/8
Randomized Algorithms, Lecture 3
47
Yao’s interpretation
• The row player = the maximizer = the
adversary responsible for designing
malicious inputs.
• The column player = the minimizer = the
algorithm designer responsible for
designing efficient algorithms.
2006/3/8
Randomized Algorithms, Lecture 3
48
Pure strategies
• For the column player (minimizer)
– Each pure strategy corresponds to a
deterministic algorithm.
• For the row player (maximizer)
– Each pure strategy corresponds to a
particular input instance.
2006/3/8
Randomized Algorithms, Lecture 3
49
Mixed strategies
• For the column player (minimizer)
– Each mixed strategy corresponds to a
randomized algorithm.
• For the row player (maximizer)
– Each mixed strategy corresponds to a
probability distribution over all the input
instances.
2006/3/8
Randomized Algorithms, Lecture 3
50
Yao’s interpretation for Loomis’
Theorem
Let T (I; A) denote the time required for algorithm A to run on
input I. Then by Loomis Theorem, we have
max
p
min
deterministic algorithm A
E[T (Ip ; A)] = min max E[T (I; Aq )]:
q
input I
Therefore, the following inequality holds for any probability distribution p and q:
min
deterministic algorithm A
2006/3/8
E[T (Ip ; A)] · max E[T (I; Aq )]:
Randomized Algorithms, Lecture 3
input I
51
Yao’s Inequality
Let ¦ be a problem. Let D consist of all deterministic algorithms for solving ¦. Let I consist of all
possible inputs for ¦. Let Ip denote the instance of
¦ with probability distribution p. Then, the following equality holds for any probability distribution p
and any Las Vegas randomized algorithm R.
·
2006/3/8
min
expected time of A running on Ip
2D
A
max
expected time of R running on I:
2I
I
Randomized Algorithms, Lecture 3
52
A comment
• The two different topics “probabilistic
analysis for deterministic algorithms” and
“randomized algorithms” interact by Yao’s
Principle.
2006/3/8
Randomized Algorithms, Lecture 3
53
How to use Yao’s Lemma?
• Task 1:
– Design a probability distribution p for the input
instance.
• Task 2:
– Obtain a lower bound on the expected
running for any deterministic algorithm
running on Ip.
2006/3/8
Randomized Algorithms, Lecture 3
54
牛刀小試
on Yao’s Principle
A lower bound (n0.694) on the
expected running time of any Las
Vegas algorithms for evaluating NOR
circuits
2006/3/8
Randomized Algorithms, Lecture 3
55
Task 1: designing Ip
x1
x2
x3
x4
y
x5
x6
x7
x8
2006/3/8
Randomized Algorithms, Lecture 3
56
An Ip
Let each leaf of the NOR circuit be independently assigned value 1 with probability
3 ¡ p5
p=
2
and value 0 with probability 1 ¡ p.
2006/3/8
Randomized Algorithms, Lecture 3
57
Interestingly,
the output of each NOR gate also has
value 1 with probability
(1 ¡ p)2 = p:
(As a matter of fact, the above equation
is how we obtained the value of p in the
¯rst place.)
2006/3/8
Randomized Algorithms, Lecture 3
58
Depth-first evaluation
• Recall that we can focus on deterministic
depth-first evaluation algorithms.
• Let A be an arbitrary algorithm of this kind.
Let W(k) be the time required for A to
evaluate the circuit on Ip with 2k numbers.
• So, W(k) = W(k – 1) + (1 – p) W(k – 1),
implying W(k) = ((2 – p)k) = (n0.694).
2006/3/8
Randomized Algorithms, Lecture 3
59
A comment
• Saks and Wigderson proved that
RandEval is optimal by designing a much
complicated Ip.
2006/3/8
Randomized Algorithms, Lecture 3
60
向左走 向右走
• The two different topics “probabilistic
analysis for deterministic algorithms” and
“randomized algorithms” meet at Yao’s
Principle.
2006/3/8
Randomized Algorithms, Lecture 3
61
© Copyright 2026 Paperzz