A Little Bit of Computational Complexity
蕭梓宏
Some interesting questions
- P=?NP
- Can we generate mathematical proofs automatically?
- Can we verify a proof looking only at very few locations of the proof?
- Does random speed up computation?
- Is approximation easier?
- How powerful is quantum computation?
Computation and Turing Machine
Turing Machine
A (k-tape) turing machine consists of:
1. k tapes that contain symbols(one of them is the input tape)
2. k read-write heads on each tape(the one on the input head is read-only)
3. a register that keeps the state of the TM
4. a finite table of instructions that given the current state and the symbols reading: write a symbol under each read/write head, move the read-write heads , and change the state
At first, the TM is in the starting state, and the input tape contains the input. Then, the TM moves according to the table of instructions until it reaches the halting state.
Turing Machine
Formally, a (k-tape) turing machine M is defined by a tuple(Γ,Q,δ):
Γ is a finite set, containing symbols that M's tapes can contain
Q is a finite set, containing the states that M can be in
δ is a function from Q×Γk to Q×Γk−1×{L,S,R}k, and is called the "transition function"
We normally assume that Γ={BLANK,START,0,1}, and Q contains qstart and qhalt
Computing a function/Running TIme
Given a function f:{0,1}∗→{0,1}∗ and another function T:N→N,
we say a TM M computes f, if for every x∈{0,1}∗, whenever M is initialized to the start configuration on input x, then it halts with f(x) written on its output tape.
We say M computes f in T(n)-time, if its computation on every input x requires at most T(∣x∣) steps.
Some useful facts about TM
- There exists an equivalent TM for any (C,java,...) program.(Operations in the machine language can be simulated by TMs)
- If f is computable on a k-tape TM in T(n)- time, then f is computable in time 5kT(n)2 on a single-tape TM whose head movements don't depend on the input and only depends on the input length.
- Each string represents a TM, and each TM can be represented by infinitely many strings
Universal TM
There exsists TM U such that for any x,α∈{0,1}∗ ,U(x,α)=Mα(x), where Mα is the TM represented by α.
Moreover, if Mα(x) halts within T steps then U(x,α) halts within CTlogT steps, where C is independent of ∣x∣
CT2 instead of CTlogT:

Universal TM with Time Bound
There exsists TM U such that for any x,α∈{0,1}∗ , T∈N,U(x,α,T)=Mα(x) if Mα(x) halts in T steps
Uncomputability/Undecidability?
演算法可數,問題不可勝數?
Uncomputable function
Define the function UC:{0,1}∗→{0,1}:
if Mα(α)=1,UC(α)=0;otherwise, UC(α)=1
No turing machine can compute UC!!!

Little fact: check if integer coefficient polynomial equations have integer solution is also uncomputable
Halting problem
HALT:{0,1}∗→{0,1}:
HALT(⟨α,x⟩)=1, if Mα(x) halts;
HALT(⟨α,x⟩)=0, otherwise
HALT is uncomputable. WHY?
HALT:{0,1}∗→{0,1}:
HALT(⟨α,x⟩)=1, if Mα(x) halts;
HALT(α)=0, otherwise
If we have Mhalt, then we can construct MUC:
1. Run Mhalt(α,α) to see if Mα(α) halts
2. If it does, run U(α,α) to get its output
3. Ouput 0 if Mα(α) halts and outputs 1; output 1 otherwise
We can REDUCE(∞⋆) UC to HALT!
UC:{0,1}∗→{0,1}:
UC(α)=0, if Mα(α)=1;
UC(α)=1, otherwise
Uncomputability and
Gödel's Imcompleteness Theorem?
Gödel's Imcompleteness Theorem
First Incompleteness Theorem: "任何可以表達算術且一致的形式系統一定是不完備的。"
一致(consistent): There is no statement ϕ such that both ϕ and ¬ϕ can be proven
完備(complete): For every statement ϕ, either ϕ or ¬ϕ can be proven
Gödel's Imcompleteness Theorem
There are true statements that cannot be proven!
(in certain axiomatic systems, like ZFC)
For example, the consistency of ZFC cannot be proven in ZFC.
Gödel's imcompleteness theorem motivated Church and Turing into their work on uncomputability.
The other way?
From Halting Problem to
Incompleteness Thorem
Weaker version of First Incompleteness Theorem: "任何可以表達算術且可靠的形式系統一定是不完備的。"
可靠(sound): If a statement ϕ is not true, it cannot be proven(which implies consistent)
完備(complete): For every statement ϕ, either ϕ or ¬ϕ can be proven
From Halting Problem to
Incompleteness Thorem
If a axiomatic system(that is powerful enough) is both sound and complete , we can devise an algorithm(or, a TM) that can compute HALT!
From Halting Problem to
Incompleteness Thorem
step 0: Read the input α and x
step 1: Express the statement "Mα halts on x" by ϕ
step 2: Enumerate all strings of finite length, and check if it is a proof for ϕ or ¬ϕ(proof checking is not hard!)
step 3: By completeness, step 2 can find the proof, and by soundness, the proven result is true. Thus we can output 1 if ϕ is true and output 0 otherwise!
Why Turing Machine?
The Church-Turing(CT) Thesis: Every effectively calculable function is a computable function
λ-calculus
general recursive function
Markov alogorithm
pointer machine
combinatory logic
computationally equivalent to TMs
P
Language/Decision Problem
A language L is a subset of
{0,1}∗(the set of all (bit)strings)
A decision problem for L is:
"Given x: output 1 if x∈L, else output 0"
*We can turn many objects into strings, like integers, pairs of integers, matrix, graph,...,and turing machines!
Language/Decision Problem
Examples of languages:
PAL:{s:s is a palindrome}
PRIME:{n:n is a prime number}
MAXINDSET:{⟨G,k⟩: the size of the largest independent set in G is k}
Complexity Class
A complexity class is a set of languages(boolean functions) that can be decided(computed) within given resource bounds.
DTIME and P
DTIME(T(n)): a language L is in DTIME(T(n)) if there is a TM that decides L in O(T(n)) time
P: P=∪c≥1DTIME(nc)
We say that problems in P are efficiently solvable
Examples?
NP
NP
a language L is in NP :
there exists a polynomial p, a polynomial-time TM M, such that ∀x∈{0,1}∗,x∈L⇔∃u∈{0,1}∣p(∣x∣)∣,s.t.M(u,x)=1
"A 'yes' instance can be verified in polynomial time with a certificate."
NP
Examples of languages in NP:
PAL:{s:s is a palindrome}
COMPOSITE:{n:n is a composite number}
SUBSETSUM:{⟨A1,A2,...,An,T⟩: there is a subset of A that sum to T}
Are they in NP?
INDSET:{⟨G,k⟩: there is a independent set of size k in G}
MAXINDSET:{⟨G,k⟩: the size of the largest independent set in G is k}
GI:{⟨G,G′⟩:G is isomorphic to G′}
GNI:{⟨G,G′⟩:G is not isomorphic to G′}
?
?
Non-Deterministic TM
An NDTM M is almost the same as a TM, but it has a special state qaccept and two transition functions δ0 and δ1. In each step of computation, M arbitrarily choose one of them to apply.
For x∈{0,1}∗, we say that M(x)=1 if there exists some sequence of these choices that makes M reach qaccept.(Otherwise, M(x) = 0)
We say that M runs in T(n) time, if for all x∈{0,1}∗ and every sequence of choices, M reaches either qhalt or qaccept within T(∣x∣) steps.
NTIME
NTIME(T(n)): a language L is in NTIME(T(n)) if there is a O(T(n)) NDTM M, such that x∈L⇔M(x)=1
Theorem 1:
NP=∪c≥1NTIME(nc)
Idea:
The choices that make the NDTM reach qaccept can be used as the certificate(and vice versa).
Polynomial-time Reduction
A language L is polynomial-time reducible to another language L′(L≤pL′), if there exists a polynomial-time computable function f:{0,1}∗→{0,1}∗ such that x∈L⇔f(x)∈L′
L≤pL′,L′≤pL′′⇒L≤pL′′
example: we can reduce 3-coloring to 4-coloring.
NP hard and NP complete
NP-hard: L′ is in NP-hard, if L≤pL′ for every L in NP
NP-complete: L′ is NP-complete if it is in NP and NP-hard
CNF(CNF formula)
a CNF instance with 5 variables and 4 clauses:
ϕ(a1,a2,a3,a4,a5)=(a1∨¬a2∨a3∨a4)∧(¬a1∨a2∨a3∨a5)∧(¬a1∨¬a4∨a5)∧(a2∨¬a3∨¬a5)
a 3CNF instance with 5 variables and 4 clauses:
ϕ(a1,a2,a3,a4,a5)=(a1∨¬a2∨a3)∧(¬a1∨a2∨a3)∧(¬a1∨¬a4∨a5)∧(a2∨¬a3∨¬a5)
A CNF is satisfiable if there exists some assignment of the variables such that ϕ is true.
clause
literal
SAT and 3SAT
SAT:{ϕ:ϕ is a satisfiable CNF}
3SAT:{ϕ:ϕ is a satisfiable 3CNF}
Cook-Levin Theorem:
SAT and 3SAT is NP-complete.
We can see that SAT and 3SAT is in NP.
Our goal is to
1. reduce every L∈NP to SAT,
2. reduce SAT to 3SAT
useful tool
Every boolean function f:{0,1}l→{0,1} can be expressed as a CNF with 2l clauses of l literals
ϕ(x)=z:f(z)=0⋀(x=z)
x_1 | x_2 | x_3 | f(x) |
---|---|---|---|
0 | 0 | 0 | 1 |
0 | 0 | 1 | 0 |
0 | 1 | 0 | 1 |
0 | 1 | 1 | 0 |
1 | 0 | 0 | 0 |
1 | 0 | 1 | 1 |
1 | 1 | 0 | 1 |
1 | 1 | 1 | 1 |
ϕ(x)=(x1∨x2∨¬x3)∧(x1∨¬x2∨¬x3)∧(¬x1∨x2∨x3)
CL Theorem Proof
We want to reduce a language L in NP to SAT(in polynomial time).
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
Since L is in NP, there is a polynomial time TM M such that
x∈L⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
CL Theorem Proof
We can assume that:
M has only two tapes (input + work / output) and it is oblivious(so the location of its heads in the i-th step only depends on the input length and i)
CL Theorem Proof
Snapshot:
The snapshot of the i-th step of M's computation on input y=x∘u is a triple ⟨a,b,q⟩∈Γ×Γ×Q, representing the symbols on the two tapes and the state that M is in.
Encode the snapshot of the i-th step as zi∈{0,1}c, where c is only dependent of ∣Γ∣ and ∣Q∣.
If someone claims that M(x,u)=1 and provides you the snapshot of the computation, how can you tell whether the snapshots present a valid computution by M?
CL Theorem Proof
zi depends on: the state in step i−1 and the symbols in the currenct cells of the two tapes
There is a function F:{0,1}2c+1→{0,1}c, such that
zi=F(zi−1,zprev(i),yinputpos(i))
F only depends on M; prev(i) and inputpos(i) only depend on M and ∣y∣, and can be known by simulating M(0∣y∣).

CL Theorem Proof
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
M(x,u)=1 for some u∈{0,1}p(∣x∣)⇔
there exist y∈{0,1}∣x∣+p(∣x∣) and z∈{0,1}c(T(n)+1), such that
1. The first n bits of y are equal to x
2. z0 encodes the initial snapshot of M
3. For every i∈[1,T(n)],zi=F(zi−1,zprev(i),yinputpos(i))
4. zT(n) encodes a snapshot in which M halts and outputs 1.
CL Theorem Proof
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
M(x,u)=1 for some u∈{0,1}p(∣x∣)⇔
there exist y∈{0,1}∣x∣+p(∣x∣) and z∈{0,1}c(T(n)+1), such that
1. "usefull tool": n clauses of size 4
2. z0 encodes the initial snapshot of M
3. For every i∈[1,T(n)],zi=F(zi−1,zprev(i),yinputpos(i))
4. zT(n) encodes a snapshot in which M halts and outputs 1.
CL Theorem Proof
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
M(x,u)=1 for some u∈{0,1}p(∣x∣)⇔
there exist y∈{0,1}∣x∣+p(∣x∣) and z∈{0,1}c(T(n)+1), such that
1. "usefull tool": n clauses of size 4
2. "useful tool": 2c clauses of size c
3. For every i∈[1,T(n)],zi=F(zi−1,zprev(i),yinputpos(i))
4. zT(n) encodes a snapshot in which M halts and outputs 1.
CL Theorem Proof
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
M(x,u)=1 for some u∈{0,1}p(∣x∣)⇔
there exist y∈{0,1}∣x∣+p(∣x∣) and z∈{0,1}c(T(n)+1), such that
1. "usefull tool": n clauses of size 4
2. "useful tool": 2c clauses of size c
3. "useful tool": T(n)23c+1 clauses of size 3c+1
4. zT(n) encodes a snapshot in which M halts and outputs 1.
CL Theorem Proof
Given x∈{0,1}∗, we want to construct ϕx such that
ϕx∈SAT⇔M(x,u)=1 for some u∈{0,1}p(∣x∣)
M(x,u)=1 for some u∈{0,1}p(∣x∣)⇔
there exist y∈{0,1}∣x∣+p(∣x∣) and z∈{0,1}c(T(n)+1), such that
1. "usefull tool": n clauses of size 4
2. "useful tool": 2c clauses of size c
3. "useful tool": T(n)23c+1 clauses of size 3c+1
4. "useful tool": 2c clauses of size c
Just let ϕx be the AND of the above conditions!
NP completeness of 3SAT
Now we know SAT is NP-complete, so we can show that 3SAT is NP-complete by reducing SAT to 3SAT in polynomial time.
(a1∨¬a2∨a3∨a4∨¬a5)⇔(a1∨¬a2∨a3∨b1)∧(a4∨¬a5∨¬b1)
⇔(a1∨¬a2∨b2)∧(a3∨b1∨¬b2)∧(a4∨¬a5∨¬b1)
We can easily turn a clause of k variables into AND of k−2 clauses of 3 variables!
Reduction 大賽
INDSET: Decide if there is a independent set of size k on G
QUADEQ: Decide if the set of boolean quadratic equations (∑i,j∈[n]ai,juiuj=b) is satisfiable(addition is modulo 2)
dHAMPATH: Decide if the directed graph G has a hamiltonian path
Exactly One 3SAT: Decide if there is an assigment such that in each clause of the CNF, exactly one of the literals is TRUE
TSP: Decide if there is a closed circuit that visits every node exactly once with total length at most k on a weighted complete graph G
HAMPATH: Decide if the undirected graph G has a hamiltonian path
SUBSETSUM: Decide if a subset of the sequence A sum up to T
coNP(順便講)
1. coNP = {L:Lˉ∈NP}
2. We say a language L is in coNP if there are a polynomial time TM M and a polynomial p such that
x∈L⇔∀u∈{0,1}p(∣x∣),M(x,u)=1
Why are the two defenitions equivalent?
example:GNI,tautology
Hardness of Approximation
MAX-3SAT
For a 3CNF ϕ, we denote val(ϕ) as the maximum fraction of clauses that can be satisfied by any assigment.
If ϕ is satisfiable, then val(ϕ)=1.
ρ-approximation
For a CNF ϕ, can we find an assignment that satisfies ρ⋅val(ϕ) of ϕ′s clauses in polynomial time?
For a graph G whose maximum independent set has size k, can we find an independent set of size ρ⋅k in polynomial time?
For a graph G whose minimum vertex cover has size k, can we find an vertex cover of size ρk in polynomial time?
Can ρ be arbitrarily close to 1?
PCP Theorem
PCP verifier?
Example: GNI∈PCP(poly(n),1)(GNI has a PCP verifier that uses poly(n) random coins and O(1) queries to the proof)
Verifier(V): By tosing random coins and making queries to the proof, the verifier should accept if G1≡G2.
Proof(π): The proof is a string, and is expected that for each n-vertex labeled graph H, π[H]=0 if H≡G0 and π[H]=1 otherwise.
V flips a coin to get b∈{0,1} and flips O(nlogn) coins to get a random permutation P of length n.
V applies P to the vertices of Gb to obtain H, and accept iff π[H]=b.
Efficient: runs in polynomial time and uses limited random coins and queries
Complete: always accepts if x∈L
Sound: accepts with probability at most 21 if x∈L
香草口味PCP theorem
NP=PCP(logn,1)
Each language in NP has a PCP verifier that uses O(logn) random coins and makes O(1) queries to the proof.
We can then probabilisticlly check a mathemaical proof by only examining a constant bit of the proof!
巧克力口味PCP theorem
There exists ρ<1, such that for every L∈NP there is a polynomial function f mapping strings to 3CNF such that :
x∈L⇒val(f(x))=1
x∈L⇒val(f(x))<ρ
Thus, there exists some ρ<1 such that if there is a ρ-approximation algortihm for MAX-3SAT, then P=NP.
Why are they equivalent?
巧克力:There exists ρ<1, such that for every L∈NP there is a polynomial function f mapping strings to 3CNF such that :
x∈L⇒val(f(x))=1
x∈L⇒val(f(x))<ρ
香草:Each language in NP has a PCP verifier V that uses O(logn) random coins and makes O(1) queries to the proof.
香草到巧克力:Define Vx,r∈{0,1} as the output of V if the random coin tosses are r(r∈{0,1}clogn). Since Vx,r only depends on q bits of the proof, we can express Vx,r by a 3CNF with q2q clauses(poly(n)⋅q2q clauses in total).
Why are they equivalent?
巧克力:There exists ρ<1, such that for every L∈NP there is a polynomial function f mapping strings to 3CNF such that :
x∈L⇒val(f(x))=1
x∈L⇒val(f(x))<ρ
香草:Each language in NP has a PCP verifier V that uses O(logn) random coins and makes O(1) queries to the proof.
巧克力到香草:For an input x, the verifier computes f(x) and expect the proof to be a satisfying assigment for f(x). Then the verifier just randomly choose one clause of f(x) and look at the corresponding three bits of the proof to check.
Maxmimum Independent Set and Minimum Vertex Cover
By the reduction of 3SAT to INDSET, we can transform a CNF ϕ to a n-vertex graph whose largest independent set has size val(ϕ)7n
By the PCP theorem, if P=NP, there exists ρ<1 such that we cannot decide if val(ϕ)=1 or val(ϕ)<ρ.
val(ϕ)
最大獨立集
最小點覆蓋
1
<ρ
7n
<ρ7n
n−7n
>n−ρ7n
n−ρ7nn−7n=7−ρ6, thus we cannot ρ-approximate the maximum independent set or 7−ρ6 -approximate the minimum vertex cover.
Approximation of maximum independent set
In fact, we cannot ρ′-approximate the maximum independent set for any ρ′!
Gk


G
G2
IS(Gk)=(IS(G))k
Maxmimum Independent Set and Minimum Vertex Cover
By the reduction of 3SAT to INDSET, we can transform a CNF ϕ to a n-vertex graph whose largest independent set has size val(ϕ)7n
val(ϕ)
IS(G)
1
<ρ
7n
<ρ7n
(7n)k
IS(Gk)
<ρk(7n)k
For any ρ′>0, there exits k such that ρk<ρ′, thus we cannot ρ′-approximate IS in polynomial time if P=NP
Proof of the PCP theorm
沒了啦><
Reference
Sanjeev Arora and Boaz Barak(2009):Computational Complexity: A Modern Approach
Communication Complexity?
What is communication complexity?
Communication Complexity
For a function f(x,y), where x∈X,y∈Y, we give x to Alice(A) and y to Bob(B), and they both need to know f(x,y).
A protocol is an algorithm that decide how A and B communicate. The length of a protocol is the maximum number of bits A and B need to exchange over all (x,y).
The communication complexity of f is the minimum length of all protocols that "computes" f.
Examples!
For x,y∈{0,1}n,
EQ(x,y)={10(if x=y)(if x=y)
Let x,y be lists of numbers in [n] whose length is t.
MD(x,y) is the median of the combination of the two lists.
* However, we usually assume f(x,y)∈{0,1}.
Formal Definition of Protocols
A protocol π can be specified by a full binary tree, where each non-leaf vertex has an owner and an associated function fv:X(or Y) →{0,1}, and each leaf has an output.
Each input (x,y) induce a path from the root to the leaf π(x,y). We say π computes a function f if the output of π(x,y) is equal to f(x,y) for all possible inputs.
The length of the protocol is the depth of the tree.

Randomized Protocols?
In the randomized case, A and B can both see an infinitely-long sequence R of unifromly random bits(independent of their inputs), and the protocol can make errors.
We say a protocol π computes f with error ϵ<21, if for all (x,y), RPr[πR(x,y)=f(x,y)]≤ϵ
The length of π is the maximum number of bits exchanged(over all R,x,y). The randomized communication complexity of f,ϵ is the minimum length of all randomized ptotocol that computes f with error ϵ.
Examples!
For x,y∈{0,1}n,
EQ(x,y)={10(if x=y)(if x=y)
Some Other Problems
For x,y∈{0,1}n,
GT(x,y)={10(if x≥y)(if x<y)
For x,y∈{0,1}n,
DISJ(x,y)={10(if ∣x∩y∣=0)(otherwise)
For x,y∈{0,1}n,
IP(x,y)=⟨x,y⟩%2
Lower Bounds
Deterministic
Functions and Matrices
For a function f(x,y), we can represent it as a matrix Mf.

MEQ(2-bit)
Rectangles
A rectangle is a subset of the input space X×Y of the form A×B, where A∈X,B∈Y.

Rectangles and Protocols
After Alice and Bob communicate according to the protocol, a rectangle is induced.
The rectangle must be monochromatic in Mf.

Rectangles and Protocols
Thus, if there exists a protocol of length c for computing f, then Mf can be partitioned into 2c monochromatic rectangles.
If every partition of Mf into monochromatic rectangles requires at least t rectangles, then the deterministic complexity of f is at least logt.
⟹
Rectangles and Protocols
Thus, if there exists a protocol of length c for computing f, then Mf can be partitioned into 2c monochromatic rectangles.
If every cover of Mf by monochromatic rectangles requires at least t rectangles, then the deterministic complexity of f is at least logt.
(∞⋆)
⟹
Rectangles and Protocols
Is the converse true?

However, if Mf can be partitioned into 2c rectangles, then there is a protocol of length O(c2) that computes f.
Deterministic Lower Bound of EQ and GT


MEQ
MGT
Fooling Set
A fooling set F of f(x,y) is a subset of inputs such that
1. f is constant on F, and
2. for all (x1,y1),(x2,y2)∈F, either (x1,y2) or (x2,y1) has the opposite f−value
The size of the (monochromatic) rectangle cover of Mf is at least the size of ∣F∣.
The deterministic communication complexity of f is at least logf.
Deterministic Lower Bound of DISJ
For x,y∈{0,1}n,
DISJ(x,y)={10(if ∣x∩y∣=0)(otherwise)
F={(x,xˉ)∣x∈{0,1}n} is a fooling set.
The Deterministic Communication Complexity of DISJ is n+1.
Rank
The deterministic communication complexity of f is at least log(rank(Mf)+1)
rank(Mf) is the smallest number r such that Mf can be expressed as the sum of r matrices of rank 1
Protocol of length c implies a partition of 2c rectangles.
Thus, rank(Mf)≤2c−1.
Lower Bounds from Rank


MEQ
MGT
Some Lemmas
rank(M⊗M′)=rank(M)⋅rank(M′)
rank(A)−rank(B)≤rank(A+B)≤rank(A)+rank(B)
Lower Bounds from Rank
MDISJ
MIP


Lower Bounds from Rank
MDISJ2
P=J−2MIP


Lower Bounds from Rank
MDISJn=[1110]⊗MDISJn−1
Pn=[111−1]⊗Pn−1
rank(MIPn)≥rank(Pn)−rank(Jn)
Randomized
Public VS Private Coin
private-coin protocol of length c ⟹ public-coin protocol of length c
public-coin protocol of length c ⟹ private-coin protocol of length O(c+logn)
(Newman's Theorem)
Randomized Protocol
=
Distribution of Deterministic Protocols
Distributional Complexity
von Neumann's minimax principle: For a n×m matrix M,
x∈Rn,∣∣x∣∣1=1,xi>0maxy∈Rm,∣∣y∣∣1=1,yi>0minxTMy=yminxmaxxTMy
The randomized complexity of f,ϵ is at least c
⇔
There is a distribution D of inputs (x,y) such that for all deterministic protocol π with error ϵ, the length of π on D is at least c
∞⋆ Yao's(姚期智) Principle

Discrepancy
For a set S⊆X×Y, and a distribution μ over the inputs, we say the discrepancy of f(x,y) w.r.t. S and μ is
∣Eμ[χS(x,y)(−1)f(x,y)]∣
low discrepancy for every rectangle ⟹ high communication complexity?
For a distribution μ, if the discrepancy of f w.r.t. every rectangle is at most γ, then the length of any ptotocol computing f with error ϵ is at least log(γ1−2ϵ) when inputs are drawn from μ.
Randomized Lower Bound of IP
For every rectangle R⊆X×Y,
∣EU[χR(x,y)(−1)IP(x,y)]∣≤2−n
The randomized communication complexity of IP is at least 2n−log(1−2ϵ1) (Ω(n))
Randomized Lower Bound of DISJ
Ω(n)
Ω(n)
Ω(n)
∞⋆!!!
Why study communication complexity?
Applications
VLSI Time-Space Tradeoff
Proof Complexity
Formula Lower Bounds
Pseudo-randomness...
Space Complexity of Streaming Models
Data Structures
Extension Complexity
Extension Complexity
E.R. Swart. P=NP. Report No. CIS86-02, Department of Computer and Information Science, University of Guelph, Ontario, Canada, 1986

(參考圖片,非當事論文)
Polytope

Intersection of several half-spaces(linear inequalities), and can be expressed as P:={x∣Ax≤b}.
face(generated by some supporting hyperplane)
vertex(also a face)
facet(also a face)
Example


By adding auxiliary variables, we obtain an extension of the original polytope.
extension
projection
Example


The extension complexity of a polytope P is the minimum number of facets(inequalities) achieved by any extension of P.
extension
projection
What is xc(TSP(n))?
Nondeterministic Communication Complexity
The setting is the same, A gets x and B gets y. However, a person C knows both x and y, and C needs to convince A and B that f(x,y)=1.
The nondeterministic communication complexity of NEQ is O(logn).
The nondeterministic communication complexity of EQ is Ω(n).
1-rectangle cover
If there is a 1-rectangle cover of size t of Mf, then there is a nondeterministic protocol of size logt.
If there is a nondeterministic protocol of length c for f, then Mf can be covered by 2c rectangles.
What is the nondeterministic communication complexity of DISJ?
Extension and Nondeterministic Communication?
xc(P)=r
rank+(S)=r, where S is the slack matrix of P.
The nondeterministic communication complexity of Face-Vertex(P) is O(logr).
⟹
Yannakakis' Factorization Theorem
⇔
Slack Matrix

A slack matrix S of P is a ∣F∣×∣V∣ matrix, where F is the set of faces, and V is the set of vertices.
Sfv=qf−pfTxv, where pfTx≤qf is the supporting hyperplane of the f-th face.
Yannakakis' Factorization Theorem-1
xc(P)≤rank+(S)
Suppose rank+(S)=r, then we can write S=RT, where R∣F∣×r,Tr×∣V∣ are nonegative.
We collect the supporting hyperplanes for every face of P, and obtain A′x≤b′, which is in fact P itself, and Si =b′−A′vi.
Let Q:={(x,y)∣A′x+Ry=b′,y≥0}, which has only r inequalities, and projx(Q)⊆P.
Also P⊆projx(Q), because for every vertex vi of P, (vi,Ti)∈Q.
Yannakakis' Factorization Theorem-2
xc(P)≥rank+(S)
Suppose xc(P)=r, and Q:={(x,y)∣Cx+Dy≤d} is an extension of P:{x∣Ax≤b}.
Farkas Lemma: Either Ax≤b has a solution x with x∈Rn, or ATy=0 has a solution y with y≥0,yTb<0.
Each supporting hyperplane generating each face of Q can be expressed as some nonegative linear combination of the facets of Q.
⟹
Yannakakis' Factorization Theorem-2
xc(P)≥rank+(S)
Suppose xc(P)=r, and Q:={(x,y)∣Cx+Dy≤d} is an extension of P:{x∣Ax≤b}.
Each supporting hyperplane generating each face of P can be expressed as some nonegative linear combination of the facets of Q.
For the f-th face of P (generated by pfTx≤qf), we can find λf∈R+r, such that λfTC=pT,λfTD=0,λfTd=q.
For the v-th vertex (xv) of P, we can find (xv,yv)∈Q, and let μv=d−Cxv−Dyv∈R+r.
λfTμv=λfT(d−Cxv−Dyv)=qf−pfTxv=Sfv
Extension and Nondeterministic Communication
xc(P)=r
rank+(S)=r, where S is the slack matrix of P.
The nondeterministic communication complexity of FACE-VERTEX(P) is O(logr).
⟹
Yannakakis' Factorization Theorem
⇔
Example: Correlation Polytope
Define CORR(n) as the convex hull of {y=xTx∣x∈{0,1}n}
Can we prove the extension complexity of CORR(n)?
We should consider the nondeterministic communication complexity of FACE−VERTEX(CORR(n)).
The Lemma
For every s∈{0,1}n, there exist a face fs of CORR(n), such that ∀x∈{0,1}n
y=xTx∈fs⟺∣s∩x∣=1
((∑si=1xi)−1)2≥0 would be great, but it is not linear.
We change xi, xi2 into yii, and change 2xixj into yij+yji to obtain fs.
Reduction?
FACE-VERTEX(CORR(n))
SIZE-OF-INTERSECTION-IS-NOT-ONE(n)
UNIQUE-DISJ(n)
easier
easier
Only consider the faces mentioned in the lemma.
Only consider (x,y) such that ∣x∩y∣≤1
Nondeterministic communication complexity of UNIQUE-DISJ!
UNIQUE-DISJ

UNIQUE-DISJ(2)
Still, we want to obtain a lower bound for the size of the 1-rectangle cover of MUNIQUE−DISJ.
1-rectangle size
Lemma: Each 1-rectangle in MUNIQUE−DISJ contains at most 2n 1-inputs.
Induction hypothesis: For a 1-rectangle R=A×B, where all x∈A,y∈B have 0s in their last n−k coordinates, R contains at most 2k 1-inputs.
k=0 is obvious, and k=n is what we want.
1-rectangle size
Induction hypothesis: For a 1-rectangle R=A×B, where all x∈A,y∈B have 0s in their last n−k coordinates, R contains at most 2k 1-inputs.
For k>0, given R, we want to construct two sets of 1-inputs S1,S2, such that
1. rect(S1),rect(S2) are 1-rectangles and have 0s in their last n−k+1 coordinates.
2. the number of 1-entries in S1,S2 is at least the number of 1-inputs in R.
We can get rid of of the last n−k coordinates.
1-rectangle size
1. rect(S1),rect(S2) are 1-rectangles and have 0s in their last n−k+1 coordinates.
2. The number of 1-entries in S1,S2 is at least the number of 1-inputs in R.
Note that (a1,b1) is not an 1-input, and (a0,b1),(a1,b0) will not be in R at the same time if (a0,b0) is an 1-input.
For every 1-input in R,
(a0,b1)⟹ put it in S1′
(a1,b0)⟹ put it in S2′
(a0,b0)⟹ put it in S1′ if (a0,b1)∈/R, put it in S2′ if (a1,b0)∈/R
rect(Si′)s are 1-rectangles.
xk=0 in rect(S1′), yk=0 in rect(S2′)
1-rectangle size
Obtain Si by setting yk=0 in rect(S1′) and xk=0 in rect(S2′)
rect(Si′)s are 1-rectangles.
xk=0 in rect(S1′), yk=0 in rect(S2′)
∣Si∣=∣Si′∣
1. rect(S1),rect(S2) are 1-rectangles and have 0s in their last n−k+1 coordinates.
2. The number of 1-entries in S1,S2 is at least the number of 1-inputs in R.
rect(Si) have 0s in their last n−k+1 coordinates.
rect(Si)s are 1-rectangles.
Proof Complete!
1-rectangle cover
Lemma: Each 1-rectangle in MUNIQUE−DISJ(n) contains at most 2n 1-inputs.
Besides, we know that there are 3n 1-inputs in MUNIQUE−DISJ(n).
Thus, any 1-rectangle cover has size at least (1.5)n, and the nondeterministic communication complexity of UNIQUE-DISJ(n) would be Ω(n).
So, finally,
xc(CORR(n))=2Ω(n)
How about TSP(n)?
TSP(n)= conv{χF∈{0,1}∣E(Kn)∣∣F⊆E(Kn)is a tour of Kn}
There is a face of TSP(cn2) that is an extension of CORR(n).
{xTx∣x∈{0,1}n}
The satisfying assingments of ϕn:=∧i,j∈[n],i=j(Cij=Cii∧Cjj)
(O(n2) variables and clauses)
Halmiltonian cycle of some directed graph Dn
(O(n2) vertices and edges)
Halmiltonian cycle of some undirected graph Gn
(O(n2) vertices and edges)
⟺
⟺
⟺
There is a face of TSP(cn2) that is an extension of CORR(n).
Extension Complexity of TSP(n)
If Q is an extension of P, then xc(Q)≥xc(P)
If P is a face of Q, then xc(Q)≥xc(P)
(Consider the slack matrix.)
There is a face of TSP(cn2) that is an extension of CORR(n).
The extension complexity of TSP(n) is 2Ω(n) ><
Q
References
[1] T. Roughgarden. Communication complexity (for algorithm designers). Foundations and Trends in Theoretical Computer Science, 11(3–4):217–404, 2016.
[2] Mathematics and Computation: A Theory Revolutionizing Technology and Science (Princeton University Press, October 2019)
[3] Rao, A., & Yehudayoff, A. (2020). Communication Complexity: And Applications. Cambridge: Cambridge University Press. doi:10.1017/9781108671644
[4] Samuel Fiorini, Serge Massar, Sebastian Pokutta, Hans Raj Tiwary, and Ronald de Wolf. 2015. Exponential Lower Bounds for Polytopes in Combinatorial Optimization. J. ACM 62, 2, Article 17 (May 2015), 23 pages. https://doi.org/10.1145/2716307
Kolmogorov Complexity and the Incompressibility Method
What's the next number?
1,2,4,8,16,
What's the next number?
1,2,4,8,16,
7122?
32?
What is randomness?
1,2,4,8,16,7122
1,2,4,8,16,32
k = 1 for i in range(5): print(k) k *= 2
k = 1 for i in range(4): print(k) k *= 2 print(7122)
A = [1,2,4,8,16,7122] for i in range(5): print(A[i])
Kolmogorov Complexity
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
Kolmogorov Complexity
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
A more formal definition
Cf(x)=min{l(p):f(p)=x}
Where f is some (partial) function.
Kolmogorov Complexity
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
A more formal definition
Cf(x)=min{l(p):f(p)=x}
Where f is some (partial) function.
How should we choose f? (C++, python, Eric Xiao, ...?)
Kolmogorov Complexity
Maybe f should be the "minimal" (additively optimal) function:
∀g,x,Cf(x)≤Cg(x)+cg
Kolmogorov Complexity
Such f doesn't exist:
Let g(i)=xi, where Cf(xi)≥i
Maybe f should be the "minimal" (additively optimal) function:
∀g,x,Cf(x)≤Cg(x)+cg
Kolmogorov Complexity
Perhaps we should not consider all functions.
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Let f0 be a function computed by a universal TM (T0) with input 11..110np (with l(n) ones ) which simulates Tn on input p
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Let f0 be a function computed by a universal TM (T0) with input 11..110np (with l(n) ones ) which simulates Tn on input p
We have Cf0(x)≤Cfn(x)+2l(n)+1
f0 is additively optimal!
Kolmogorov Complexity
C(x)=Cf0(x)=min{l(p):f0(p)=x}
Kolmogorov Complexity
C(x)=Cf0(x)=min{l(p):f0(p)=x}
We can view f0 as a special compiler.
Is the compiler of C/Java additively optimal?
Conditional Kolmogorov Complexity
C(x∣y)=Cf0(x∣y)=min{l(p):f0(⟨y,p⟩)=x}
C(x)=C(x∣ϵ)
We also have ∀f,x,y
Cf0(x∣y)≤Cf(x∣y)+cf
Upper bounds
∀x,C(x)≤l(x)+c
∀x,y,C(x∣y)≤C(x)+c
Uncomputability
C(x) is uncomputable
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
Uncomputability
C(x) is uncomputable
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
「不能用二十個字以內描述的最小數字」
Uncomputability
C(x) is uncomputable on any infinite set of points
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
「不能用二十個字以內描述的最小數字」
Incompressibility Theorem ∞⋆
Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Incompressibility Theorem ∞⋆
Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Because there are only i=0∑logm−c−12i=m2−c−1 programs with length less than logm−c.
Example - Infinite Primes
Assume that there are only k primes. Then for each n∈N, we can write n=i=1∏kpiαi. Thus we can describe every number in [1,t] by kloglogt bits......
Incompressity theorem: Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Example - Infinite Primes
Assume that there are only k primes. Then for each n∈N, we can write n=i=1∏kpiαi. Thus we can describe every number in [1,t] by kloglogt bits......
Incompressity theorem: Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
π(n)∈Ω(loglognlogn)
Example - TM complexity
Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Example - TM complexity
Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Assume a TM T:
- each state can be described by ℓ bits
- always terminates with the reading head at the rightmost position of the input
- computes PALINDROME
Example - TM complexity
Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Assume a TM T:
- each state can be described by ℓ bits
- always terminates with the reading head at the rightmost position of the input
- computes PALINDROME
Define the crossing seuence of a grid g: the sequence of states of the TM T when the reading head crosses the right boundary of g
Example - TM complexity
By the incompressibility theorem, we can find length-n string x with C(x∣T,n)≥n.
Example - TM complexity
By the incompressibility theorem, we can find length-n string x with C(x∣T,n)≥n.
Consider running T on x02nxˉ, if the runtime is less than ℓn2, then there exists a "middle" grid g0 with crossing sequence shorter than 2ℓn.
In fact, given T,n, we can recover x by the position of g0(2n bits) and the crossing sequence of g0(O(logn) bits).
Example - TM complexity
By the incompressibility theorem, we can find length-n string x with C(x∣T,n)≥n.
Consider running T on x02nxˉ, if the runtime is less than ℓn2, then there exists a "middle" grid g0 with crossing sequence shorter than 2ℓn.
In fact, given T,n, we can recover x by the position of g0(2n bits) and the crossing sequence of g0(O(logn) bits).
C(x∣T,n)≤2n+O(logn)
Example - TM complexity
We can recover x by checking if the crossing sequence at g0 with input y02nxˉ is the same as x02nxˉ for all y∈{0,1}n !
Example - TM complexity
We can recover x by checking if the crossing sequence at g0 with input y02nxˉ is the same as x02nxˉ for all y∈{0,1}n !
How about average complexity?
Incompressibility Method
If the statement doesn't hold, then there will be too many compressible objects.
Combinatorial Properties
Some Useful Bounds
en−1nn≤n!≤en−1nn+1
(kn)k≤Ckn≤k!nk<(kne)k
Transitive Tournament
Tournament: complete directed graph
Transitive tournament: acyclic tournament
v(n): Largest integer such that every tournament on n nodes contains a subtournament on v(n) nodes.
Transitive Tournament
Tournament: complete directed graph
Transitive tournament: acyclic tournament
v(n): Largest integer such that every tournament on n nodes contains a subtournament on v(n) nodes.
⌊logn⌋≤v(n)≤1+⌊2logn⌋
Counting
v(n)≤1+⌊2logn⌋
v:=2+⌊2logn⌋
A: subset of [n] of size v
σ: permutation on [v]
ΓA,σ: the set of n- vertex tournaments of whom A is a subtournament with "order" σ
We count the number d(Γ′) of n- vertex tournaments with a transitive subtournament of size v
d(Γ′)≤∑A∑σd(ΓA,σ)=Cvnv!2C2n−C2v<2C2n
(Cvnv!<nv≤2C2v)
Probabilistic Method
v(n)≤1+⌊2logn⌋
v:=2+⌊2logn⌋
A: subset of [n] of size v
σ: permutation on [v]
Let T be the random variable uniformly distributed over all n-vertex tournaments.
We calculate the probability P(T∈Γ′).
P(T∈Γ′)≤∑A∑σP(T∈ΓA,σ)=Cvnv!2−C2v<1
(Cvnv!<nv≤2C2v)
Incompressibility Method
v(n)≤1+⌊2logn⌋
If v(n) is too large, then we can compress every n-vertex tournament. (contradicting the incompressibility theorem)
Formally speaking, by the incompressibility method, there is some tournament T′, where C(T′∣n,p)≥n(n−1)/2.
(p is some decoding program)
However, for every tournament T, we can describe it by v(n)⌊logn⌋+n(n−1)/2−v(n)(v(n)−1)/2 bits.
Thus v(n)⌊logn⌋−v(n)(v(n)−1)/2≥0, and
v(n)≤2⌊logn⌋+1.
Incompressibility Method
In fact, by the incompressibility method, we can also prove:
For at least a (1−n1)th fraction of n-vertex tournaments, the largest transitive subtournament has at most (1+2⌊2logn⌋) nodes for n large enough.
v(n)≤1+⌊2logn⌋
Excerises
1. Give an upper bound for w(n), the largest integer such that for each n-vertex tournament T, there are two disjoint subsets of vertices A,B, each of size w(n), and A×B∈T.
// x,y are n-bit s = (x ^ y) c = (x & y) while c != 0: s' = (s ^ c) c' = (s & c) s = s' c = c'
2. Prove that the while loop loops at most logn+1 steps in average:
4. What are the differences between the counting argument, the probabilistic method, and the incompressibility method?
5. Ask a question.
3. Prove that almost all n-vertex labeled trees have maximum degree O(loglognlogn).
W.W. Kirchherr, Kolmogorov complexity and random graphs
Time Complexity
Shell Sort
h1,h2,...,hp=1
In the i−th pass, divide the list into hi sublists (by (pmodhi)) and do insertion sort.
h=[5,3,1]:
Average Time Complexity of Shell Sort
Pratt: Θ(nlog2n) (worst case also), using hk∈{2i3j∣2i3j<2N}
Knuth: Θ(n5/3), using the best choice of h for p=2
Yao, Janson, Knuth: O(n23/15), using some choice of h for p=3
We prove that running shell sort for any p and (h1,h2,...,hp), the time complexity is Ω(pn1+1/p).
Proof Outline
Fix some permutation π such that C(π∣n,A,p)≤logn!−logn,
where A is the algorithm, and p is some decoding program.
*A fraction of (1−n1) permutations satisfy this condition.
Let mi,k be the distance that the number i moved in pass k, and M=i=1∑nk=1∑pmi,k, which is a lower bound of the runtime.
By all mi,k, we can recover π.
If M=o(pn1+1/p), we can express all mi,k with too few bits. (TBD)
Thus the runtime is Ω(pn1+1/p) for a fraction of (1−n1) permutations and so is the average runtime.
Proof Detail
To describe all mi,k, we can descibe
the sum of them (M, described by O(logn) bits), and
the way of partitioning M into np parts (described by log(HMnp) bits).
Thus log(HMnp)+O(logn)≥logn!−logn,
log(Cnp−1M+np−1)≥logn!+O(logn)
(np−1)[log(np−1M+1)+loge]≥nlogn−nloge+O(logn)
(using Ckn≤(kne)k and n!≥ennn)
Thus M=Ω(pn1+1/p) ???
Regular Language
Regular Language
A language accepted by some DFA.

Pumping Lemma

Pumping Lemma

Might not be easy to use.
Not characterizing.
Example
Is L:={0k1k∣k≥1} a regular language?
Example
Is L:={0k1k∣k≥1} a regular language?
Suppose L is accepted by some DFA A. Then we can recover k from A and q, the state of A after processing 0k. Thus for all k≥1, C(k)≤O(1).
KC-Regularity
Suppose L is regular, and Lx:={y∣xy∈L}. If y is the n-th string in Lx, then C(y)≤C(n)+O(1).
KC-Regularity
Suppose L is regular, and Lx:={y∣xy∈L}. If y is the n-th string in Lx, then C(y)≤C(n)+O(1).
*What does "the n-th string" mean?
More Examples
Is L:={1p∣p is prime} a regular language?
Is L:={xxˉw∣∣x∣,∣w∣>0} a regular language?
Regular KC-Characterization
In fact, we have
L is regular
⇔
Exists cL, such that for all x,n, C(χ1:nx)≤C(n)+cL
For all x,n, if y is the n-th string in Lx, then C(y)≤C(n)+O(1).
⇔
Regular KC-Characterization
In fact, we have
L is regular
⇔
Exists cL, such that for all x,n, C(χ1:nx)≤C(n)+cL
For all x,n, if y is the n-th string in Lx, then C(y)≤C(n)+O(1).
⇔
Thus, (in theory), we can prove non-regularity of every non-regular language and regularity of every regular language.
Last Example
Is L:={x∣∑xi=2k+1} a regular language?
Last Example
Is L:={x∣∑xi=2k+1} a regular language?
There exists cL, such that for all x,n, C(χ1:nx)≤C(n)+cL.
Excerises
1. Prove (or disprove) that L:={0i1j∣gcd(i,j)=1} is regular.
2. Alice and Bob are playing the number guessing game. Alice chooses some x∈[n], and Bob can ask her yes/no questions. If Alice can lie once, what is the minimum number c where Bob can always guess the right answer after asking c questions?
5. Why is computability important?
6. Ask a question.
4. Prove any (nontrivial?) statement via the incompressibility method.
3. An nni (nearest-neighbor interchange) operation can be done on neighboring internal nodes on an evolutionary tree. Let f(A,B) denote the nni-distance between two n-node evolutionary trees. Prove that maxA,Bf(A,B)=Ω(nlogn). (In fact, Θ(nlogn).)

References
[1] Wikipedia contributors. (2023, March 5). Incompressibility method. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Incompressibility_method&oldid=1143081380
[2] Wikipedia contributors. (2023, August 22). Kolmogorov complexity. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Kolmogorov_complexity&oldid=1171617849
[3] Ming Li , Paul Vitányi. An Introduction to Kolmogorov Complexity and Its Applications, 2008.
A Glimpse into
Algorithmic Game Theory
2024.2.15
Other Considerd Topics - Spectral Graph Theory
- Spectrum of Adjacency Matrix/Laplacian
- Expander Graphs/PSRGs
- Expander Codes
- Nearly-linear time Laplacian solvers
Other Considerd Topics - Sum of Squares Hierachy
- Semidefinite Programming/Polynomial Optimization?
- polynomial-time 0.878-approximation algorithm for max cut (and why this might be optimal)
Today's Topic - Algorithmic Game Theory
- Auction Analysis
- Price of Anarchy (PoA, 最壞均衡與最佳解比)
Reference (mostly by Tim Roughgarden):
Barbados Lectures on Complexity Theory, Game Theory, and Economics

What is a Game?
A game is defined by
1. A set of n players
2. The pure strategy set for each player: S1,S2,...,Sn
3. The payoff function for each player: u1,u2,...,un:S1×S2×...×Sn→R

https://www.youtube.com/watch?app=desktop&v=-1GDMXoMdaY
Equilibriums
A pure strategy profile s∗∈S1×S2×...×Sn is a Pure Nash Equilibrium (PNE) if
ui(s∗)≥ui(si,s−i∗)∀i∈[n] and si∈Si
A mixed strategy profile p∗∈Δ(S1)×Δ(S2)×...×Δ(Sn) is a Mixed Nash Equilibrium (MNE) if
Es∗∼p∗[ui(s∗)]≥Es−i∗∼p−i∗[ui(si,s−i∗)]∀i∈[n] and si∈Si
Equilibriums
A pure strategy profile s∗∈S1×S2×...×Sn is a Pure Nash Equilibrium (PNE) if
ui(s∗)≥ui(si,s−i∗)∀i∈[n] and si∈Si
A mixed strategy profile p∗∈Δ(S1)×Δ(S2)×...×Δ(Sn) is a Mixed Nash Equilibrium (MNE) if
Es∗∼p∗[ui(s∗)]≥Es−i∗∼p−i∗[ui(si,s−i∗)]∀i∈[n] and si∈Si
Nash's theorem: There is at least one MNE for each game.
Q: What is an MNE for two-player rock paper scissors?
Equilibriums
A pure strategy profile s∗∈S1×S2×...×Sn is a Pure Nash Equilibrium (PNE) if
ui(s∗)≥ui(si,s−i∗)∀i∈[n] and si∈Si
A mixed strategy profile p∗∈Δ(S1)×Δ(S2)×...×Δ(Sn) is a Mixed Nash Equilibrium (MNE) if
Es∗∼p∗[ui(s∗)]≥Es−i∗∼p−i∗[ui(si,s−i∗)]∀i∈[n] and si∈Si
Nash's theorem: There is at least one MNE for each game.
Q: Why are equilibriums important?
What is Economics?
What is Economics?
- 經世濟民
- 尋找最有效率分配資源的方式
An Efficient Auction Mechanism:
Single-Item Second-Price Auction
1. Each of the n bidders has a private valuation vi of the item.
2. Each bidder submits its bid bi to the auctioneer simultaneously.
3. The auctioneer gives the item to the k-th bidder where bk=maxibi and charges p(b)=maxj=kbj
An Efficient Auction Mechanism:
Single-Item Second-Price Auction
1. Each of the n bidders has a private valuation vi of the item.
2. Each bidder submits its bid bi to the auctioneer simultaneously.
3. The auctioneer gives the item to the k-th bidder where bk=maxibi and charges p(b)=maxj=kbj
- Bidding truthfully (bi=vi) maximizes the utility ui(b)=(vi−p(b))⋅[i=k] for every bidder.
- The social welfare SW(b)=vk is always maximized when all bidders bid truthfully.
How about Single-Item First-Price Auctions?
1. Each of the n bidders has a private valuation vi of the item.
2. Each bidder submits its bid bi to the auctioneer simultaneously.
3. The auctioneer gives the item to the k-th bidder where bk=maxibi and charges bk
How about Single-Item First-Price Auctions?
1. Each of the n bidders has a private valuation vi of the item.
2. Each bidder submits its bid bi to the auctioneer simultaneously.
3. The auctioneer gives the item to the k-th bidder where bk=maxibi and charges bk
Q:What will the bidders' strategy be? Is the welfare always maximzied?
How about Single-Item First-Price Auctions?
1. Each of the n bidders has a private valuation vi of the item.
2. Each bidder submits its bid bi to the auctioneer simultaneously.
3. The auctioneer gives the item to the k-th bidder where bk=maxibi and charges bk
Questions we will consider:
- How can we define the efficiency(inefficiency) of the first-price auction?
- How can we find/prove the efficiency of the first-price auction?
- Other kinds of auctions? More than one item?
Model:
Games(Auctions) of Incomplete Information
Assumption:
The valuation vi of each bidder is drawn independently from some publicly-known distribution Fi. (Fi Corresponds to the common beliefs that bidders have about everyone’s valuations.)
A strategy of the i-th bidder is a function si:SUPP(Fi)→R+. (When my valuation is vi , I will bid si(vi).)
A strategy profile s is a Bayes-Nash Equilibrium(BNE) if ∀i∈[n],vi∈SUPP(Fi)
Ev−i∼F−i[ui(s(v−i,vi))]≥Ev−i∼F−i[ui(s−i(v−i),r′)], ∀r′∈R+
Model:
Games(Auctions) of Incomplete Information
Assumption:
The valuation vi of each bidder is drawn independently from some publicly-known distribution Fi. (Fi Corresponds to the common beliefs that bidders have about everyone’s valuations.)
A strategy of the i-th bidder is a function si:SUPP(Fi)→R+. (When my valuation is vi , I will bid si(vi).)
A strategy profile s is a Bayes-Nash Equilibrium(BNE) if ∀i∈[n],vi∈SUPP(Fi)
Ev−i∼F−i[ui(s(v−i,vi))]≥Ev−i∼F−i[ui(s−i(v−i),r′)], ∀r′∈R+
Q : The mixed version?
Example:
Two bidders, symmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1,v2 i.i.d. follow U[0,1]. Please find a (pure) BNE.
Example:
Two bidders, symmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1,v2 i.i.d. follow U[0,1]. Please find a (pure) BNE.
Hint: The two bidders' strategies are identical and monotone.
Example:
Two bidders, symmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1,v2 i.i.d. follow U[0,1]. Please find a (pure) BNE.
Q:
Is the BNE easy to solve?
Does the BNE maximize the welfare?
Is the auction efficient in this case?
Example:
Two bidders, asymmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1∼U[0,1],v2∼U[0,2] where v1⊥v2. Please find the BNE.
Example:
Two bidders, asymmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1∼U[0,1],v2∼U[0,2] where v1⊥v2.
s1(v1)=3v14(1−1−43v12)
s2(v2)=3v24(1+43v22−1)
Example:
Two bidders, asymmetric uniform valuation
Q:
There are two bidders in a first-price auction, and v1∼U[0,1],v2∼U[0,2] where v1⊥v2.
Q:
Is the BNE easy to solve?
Does the BNE maximize the welfare?
Is the auction efficient in this case?
s1(v1)=3v14(1−1−43v12)
s2(v2)=3v24(1+43v22−1)
Some notations
Valuation profile: v=(v1,v2,...,vn)
Bid profile: b=(b1,b2,...,bn)
Whether or not bidder i is the winner: xi(b)
Selling price: p(b)=maxibi
Utility: ui(b,vi)=(vi−bi)⋅xi(b)
Strategy profile: s=(s1,s2,...,sn):SUPP(F)→Rn
Social Welfare: SW(b;v)=i=1∑nvi⋅xi(b)
Whether or not bidder i has the highest valuation: xi∗(v)
Optimal social welfare: OPT(v)=maxivi=i=1∑nvi⋅xi∗(v)
∞⋆ Price of Anarchy
Social Welfare: SW(b;v)=i=1∑nvi⋅xi(b)
Optimal social welfare: OPT(v)=maxivi =i=1∑nvi⋅xi∗(v)
Price of Anarchy: PoA=mins is a BNEEv∼F[OPT(v)]Ev∼F[SW(s(v);v)]
∞⋆ Price of Anarchy
Social Welfare: SW(b;v)=i=1∑nvi⋅xi(b)
Optimal social welfare: OPT(v)=maxivi =i=1∑nvi⋅xi∗(v)
Price of Anarchy: PoA=mins is a BNEEv∼F[OPT(v)]Ev∼F[SW(s(v);v)]
High PoA ⟹ efficient auction?
Low PoA ⟹ inefficient auction?
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
ui(b−i,2vi;vi)≥(21vi−p(b))
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
ui(b−i,2vi;vi)≥(21vi−p(b))⋅xi∗(v)
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
ui(b−i,2vi;vi)≥(21vi−p(b))⋅xi∗(v)
2. For every BNE s, bidder i, and its valuation vi
Ev−i[ui(s(v);vi)]≥Ev−i[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i, and its valuation vi
Ev−i[ui(s(v);vi)]≥Ev−i[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i,
EviEv−i[ui(s(v);vi)]≥EviEv−i[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i,
Ev[ui(s(v);vi)]≥Ev[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i,
Ev∑i[ui(s(v);vi)]≥Ev∑i[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i,
EvSW(s(v);v)−p(s(v))=Ev∑i[ui(s(v);vi)]≥Ev∑i[ui(2vi,s−i(v−i);vi)].
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
Proof: We will proof the 21 PoA bound. There are two main inequalities.
1. For every valuation profile v, bid profile b, and bidder i:
∑iui(b−i,2vi;vi)≥∑i(21vi−p(b))⋅xi∗(v)=21OPT(v)−p(b)
2. For every BNE s, bidder i,
EvSW(s(v);v)−p(s(v))=Ev∑i[ui(s(v);vi)]≥Ev∑i[ui(2vi,s−i(v−i);vi)].
⟹PoA≥21
Theorem 1
The PoA of first-price single-item auction is at least 1−e1≈0.63.
* The 1−e1 bound can be proved similarly.
* The bound holds even if the valuations are correlated!
More Items? Simultaneous First-price Auctions
1. Each of the n bidders has a private valuation vij(i∈[n],j∈[m]) to each of the m item. (Assume vi∼Fi)
2. Each bidder submits its bid bi=(bi1,bi2,...,bim) to the auctioneer simultaneously.
3. The auctioneer gives the j-th item to the k-th bidder where bkj=maxibij and charges bkj
More Items? Simultaneous First-price Auctions
1. Each of the n bidders has a private valuation vij(i∈[n],j∈[m]) to each of the m item. (Assume vi∼Fi)
2. Each bidder submits its bid bi=(bi1,bi2,...,bim) to the auctioneer simultaneously.
3. The auctioneer gives the j-th item to the k-th bidder where bkj=maxibij and charges bkj
Denote by Si(b) the set of items given to bidder i.
We assume the bidders are unit-demand:
ui(b;vi)=maxj∈Si(b)vij−∑j∈Si(b)bij
More Items? Simultaneous First-price Auctions
1. Each of the n bidders has a private valuation vij(i∈[n],j∈[m]) to each of the m item. (Assume vi∼Fi)
2. Each bidder submits its bid bi=(bi1,bi2,...,bim) to the auctioneer simultaneously.
3. The auctioneer gives the j-th item to the k-th bidder where bkj=maxibij and charges bkj
Denote by Si(b) the set of items given to bidder i.
We assume the bidders are unit-demand:
ui(b;vi)=maxj∈Si(b)vij−∑j∈Si(b)bij
Maximum weight matching?
Is there a PoA bound for
S1A with unit-demand bidders?
Theorem 2
The PoA of S1A with unit demand bidders is at least 1−e1≈0.63.
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
We will first prove the complete-information version:
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Theorem 2
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
Denote by j∗(i), the item bidder i gets in some fixed optimal allocation. Define bidder i's deviation bi∗ as bidding 2vij∗(i) on j∗(i) and 0 on other items.
Theorem 2
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
Denote by j∗(i), the item bidder i gets in some fixed optimal allocation. Define bidder i's deviation bi∗ as bidding 2vij∗(i) on j∗(i) and 0 on other items.
Like the proof of theorem 1, we have for every bid profile b,
ui(bi∗,b−i;υi)≥2vij∗(i)−pj∗(i)(b)
Theorem 2
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
Like the proof of theorem 1, we have for every bid profile b,
ui(bi∗,b−i;υi)≥2vij∗(i)−pj∗(i)(b)
Thus ∑i∈[n]ui(bi∗,b−i;vi)≥21OPT(v)−∑j∈[m]pj(b).
Theorem 2
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
Like the proof of theorem 1, we have for every bid profile b,
ui(bi∗,b−i;υi)≥2vij∗(i)−pj∗(i)(b)
Thus ∑i∈[n]ui(bi∗,b−i;vi)≥21OPT(v)−∑j∈[m]pj(b).
Like the proof of theorem 1, using the definition of MNE, we can have PoA ≥21.
Theorem 2
Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi is a degenarate distribution at vi) achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
Like the proof of theorem 1, we have for every bid profile b,
ui(bi∗,b−i;υi)≥2vij∗(i)−pj∗(i)(b)
Thus ∑i∈[n]ui(bi∗,b−i;vi)≥21OPT(v)−∑j∈[m]pj(b).
Like the proof of theorem 1, using the definition of MNE, we can have PoA ≥21.
Can we use the same proof for the incomplete-information case?
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We fix some BNE, where bidder i samples its bid from the distribution Di(vi) when its valuation is vi. Denote by Gi the distribution of bidder i's bid, and G=G1×G2×...×Gn.
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We fix some BNE, where bidder i samples its bid from the distribution Di(vi) when its valuation is vi. Denote by Gi the distribution of bidder i's bid, and G=G1×G2×...×Gn.
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE, for every i and vi:
Ebi∼D(vi),b−i∼G−i[ui(b;vi)]≥ Ebi′∼D′(vi),b−i∼G−i[ui(b−i,bi′;vi)]
=Ev−i∼F−i,b−i∼G−i[ui(b−i,bi∗(v);vi)]
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE, for every i and vi:
Ebi∼D(vi),b−i∼G−i[ui(b;vi)]≥ Ev−i∼F−i,b−i∼G−i[ui(b−i,bi∗(v);vi)]
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE, for every i:
Evi∼Fi,bi∼D(vi),b−i∼G−i[ui(b;vi)]≥ Ev∼F,b∼G[ui(b−i,bi∗(v);vi)]
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE:
∑iEvi∼Fi,bi∼D(vi),b−i∼G−i[ui(b;vi)]≥ Ev∼F,b∼G[∑iui(b−i,bi∗(v);vi)]
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE:
∑iEvi∼Fi,bi∼D(vi),b−i∼G−i[ui(b;vi)]≥ Ev∼F,b∼G[∑iui(b−i,bi∗(v);vi)]
Last Proof:
∑iui(bi∗,b−i;vi)≥21OPT(v)−∑j∈[m]pj(b).
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE:
Ev∼F,bi∼D(vi)∀i[∑iui(b;vi)]≥ 21Ev∼F[OPT(v)]−Eb∼G[R(b)]
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE:
Ev∼F,bi∼D(vi)∀i[(∑iui(b;vi))+R(b)]≥ 21Ev∼F[OPT(v)]
Expected social welfare under the BNE.
Theorem 2
Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 1−e1 times the expected optimal welfare.
Proof: (We will still prove the 21 PoA bound only.)
We will use the randomized deviation bi′∼Di′(vi): sample v−i from F−i and then performs bi∗(v).
By the definition of BNE:
Ev∼F,bi∼D(vi)∀i[(∑iui(b;vi))+R(b)]≥ 21Ev∼F[OPT(v)]
Expected social welfare under the BNE.
DONE!
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
The proof techniques are similar:
1. Choose some deviation bi′.
2. Find some lower bound for ∑iui(b−i,bi′;vi) where the lower bound is related with OPT(v).
3. By the definition of BNE, the lower bound in 2. is also a lower bound for the expected social welfare under the BNE.
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
The proof techniques are similar:
Can the proof technique be used to prove PoA for other auctions?
Can we turn the proof technique into some general theorem?
General Auction Mechanisms
1. The auctioneer receives an action ai∈Ai from each player
2. Given a=(a1,a2,...,an), the auctioneer decides an outcome o(a)∈O
3. Each player derives some utility ui(a;vi):=ui(o(a);vi), where vi is the valuation of the player
4. R(a):=R(o(a)) is the revenue of the auctioneer
*The BNE, social welfare, optimal welfare, PoA can also be defined.
General Auction Mechanisms
1. The auctioneer receives an action ai∈Ai from each player
2. Given a=(a1,a2,...,an), the auctioneer decides an outcome o(a)∈O
3. Each player derives some utility ui(a;vi):=ui(o(a);vi), where vi is the valuation of the player
4. R(a):=R(o(a)) is the revenue of the auctioneer
Examples:
- combinatorial auctions: m items and n players, each player has a valuation for each of the 2m subsets of items
- single item auctions (first price, second price, all pay auctions, ...)
- S1A with unit-demand bidders
General Auction Mechanisms
1. The auctioneer receives an action ai∈Ai from each player
2. Given a=(a1,a2,...,an), the auctioneer decides an outcome o(a)∈O
3. Each player derives some utility ui(a;vi):=ui(o(a);vi), where vi is the valuation of the player
4. R(a):=R(o(a)) is the revenue of the auctioneer
Examples:
- combinatorial auctions
- public good: There is some cost c publicly known, and each of the n players has its own valuation vi.
Smooth Auction
For λ≥0,μ≥1, an auction is (λ,μ)-smooth if for every valuation profile v, there exists action distributions D1∗(v),D2∗(v),...,Dn∗(v) such that for every action profile a:
∑iEai∗∼Di∗(v)[ui(ai∗,a−i;vi)]≥λOPT(v)−μR(a).
Smooth Auction
For λ≥0,μ≥1, an auction is (λ,μ)-smooth if for every valuation profile v, there exists action distributions D1∗(v),D2∗(v),...,Dn∗(v) such that for every action profile a:
∑iEai∗∼Di∗(v)[ui(ai∗,a−i;vi)]≥λOPT(v)−μR(a).
Almost same proof of Theorem 2
(complete information version)
If an auction is (λ,μ)-smooth, then for every valuation profile v, every complete-information MNE of the auction has expected welfare at least μλ⋅OPT(v).
Smooth Auction
For λ≥0,μ≥1, an auction is (λ,μ)-smooth if for every valuation profile v, there exists action distributions D1∗(v),D2∗(v),...,Dn∗(v) such that for every action profile a:
∑iEai∗∼Di∗(v)[ui(ai∗,a−i;vi)]≥λOPT(v)−μR(a).
Almost same proof of Theorem 2
(incomplete information version)
∞⋆ If an auction is (λ,μ)-smooth, then for all valuation distributions F1,F2,...,Fn, every BNE of the auction has expected welfare at least μλ⋅Ev∼F[OPT(v)].
Excercises!
1. Prove the 21 PoA bound for single-item all-pay auction.
2. Prove the (1−e1) PoA bound for single-item first-price auction.
3*. Prove a PoA bound for the single-item auction where the two bidders with the highest bids needs to pay and the bidder with the highest bid gets the item (or prove that no PoA bound holds).
4. 剛剛講的那些東西"algorithmic"在哪裡?
5. When were the mentioned results derived?
1920 (A) 1940 (B) 1960 (C) 1980 (D) 2000 (E) 2020
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.
Critique: The strategies of the players do not always form BNE in reality!
No-Regret Learning
In this scenario, we assume that the same n bidders participate in the same auction for T rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid.
No-Regret Learning
In this scenario, we assume that the same n bidders participate in the same auction for T rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid.
We assume that all players choose their bidding stategies ait to have o(T) regret (no regret), where the regret is defined as
RT=(maxa′t=1∑Tuit(a′))−(t=1∑Tuit(ait) )
No-Regret Learning
In this scenario, we assume that the same n bidders participate in the same auction for T rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid.
We assume that all players choose their bidding stategies ait to have o(T) expected regret (no regret), where the regret is defined as
RT=(maxa′t=1∑Tuit(a′))−(t=1∑Tuit(ait) )
Q: Is there really a way to have o(T) expected regret surely?
No-Regret Learning
In this scenario, we assume that the same n bidders participate in the same auction for T rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid.
We assume that all players choose their bidding stategies ait to have o(T) expected regret (no regret), where the regret is defined as
RT=(maxa′t=1∑Tuit(a′))−(t=1∑Tuit(ait) )
Q: Is there really a way to have o(T) expected regret surely?
Yes! For example, the Multiplicative Weights Update algorithm.
For those interested, please refer to Prediction, Learning, and Games.
No-Regret Learning
If every bidder uses an no-regret learning algorithm, we get a vanishing regret sequence a1,a2,a3,... of action profiles, That is, for every i and ai′∈Ai
limT→∞T1∑t=1T(ui(ai′,a−it;vi)−ui(at;vi))≤0.
No-Regret Learning
If every bidder uses an no-regret learning algorithm, we get a vanishing regret sequence a1,a2,a3,... of action profiles, That is, for every i and ai′∈Ai
limT→∞T1∑t=1T(ui(ai′,a−it;vi)−ui(at;vi))≤0.
We have the following theorem for PoA of vanishing regret sequence:
If an auction is (λ,μ)-smooth, then for every valuation profile v, every vanishing regret sequence of the auction has expected welfare at least μλ⋅OPT(v) as T→∞.
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with unit-demand bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
Critique: How about S1A for more general bidders?
S1A for bidders with Monotone Submodular Valuations
For a set I of items, a valuation vi:2I→R is monotone submodular, if
1. For S⊆T, vi(S)≤vi(T)
2. For S⊆T and any item j, vi(T∪{j})−vi(T)≤vi(S∪{j})−vi(S)
S1A for bidders with Monotone Submodular Valuations
For a set I of items, a valuation vi:2I→R is monotone submodular, if
1. For S⊆T, vi(S)≤vi(T)
2. For S⊆T and any item j, vi(T∪{j})−vi(T)≤vi(S∪{j})−vi(S)
diminishing returns
S1A for bidders with Monotone Submodular Valuations
For a set I of items, a valuation vi:2I→R is monotone submodular, if
1. For S⊆T, vi(S)≤vi(T)
2. For S⊆T and any item j, vi(T∪{j})−vi(T)≤vi(S∪{j})−vi(S)
* Additive and unit-demand valuations are monotone submodular.
S1A for bidders with Monotone Submodular Valuations
For a set I of items, a valuation vi:2I→R is monotone submodular, if
1. For S⊆T, vi(S)≤vi(T)
2. For S⊆T and any item j, vi(T∪{j})−vi(T)≤vi(S∪{j})−vi(S).
We can prove that S1A for bidders with monotone submodular valuations is (1−e1,1) - smooth, thus the PoA bound is 1−e1.
S1A for bidders with Monotone Submodular Valuations
For a set I of items, a valuation vi:2I→R is monotone submodular, if
1. For S⊆T, vi(S)≤vi(T)
2. For S⊆T and any item j, vi(T∪{j})−vi(T)≤vi(S∪{j})−vi(S).
We can prove that S1A for bidders with monotone submodular valuations is (1−e1,1) - smooth, thus the PoA bound is 1−e1.
The bound is proven with the composition theorem:
If players have complement-free utility functions, then the simultaneous composition of (λ,μ)-smooth auctions is again a (λ,μ)-smooth auction. (See theorem 6.1, Price of Anarchy in Auctions)
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with monotone submodular bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with monotone submodular bidders is at least 1−e1
- Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
Critique: How about S1A for more general bidders?
S1A for bidders with Monotone Subadditive Valuations
For a set I of items, a valuation vi:2I→R is monotone subadditive, if
1. For S⊆T, vi(S)≤vi(T)
2. For all S,T, vi(S∪T)≤vi(S)+vi(T)
* Submodular valuations are subadditive.
S1A for bidders with Monotone Subadditive Valuations
For a set I of items, a valuation vi:2I→R is monotone subadditive, if
1. For S⊆T, vi(S)≤vi(T)
2. For all S,T, vi(S∪T)≤vi(S)+vi(T)
* Submodular valuations are subadditive.
In fact, a PoA bound of 21 can be proven for S1A for bidders with monotone subadditive valuations.
* Not proven by the smoothness argument, see Tim's lecture notes.
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with monotone submodular bidders is at least 1−e1
-
Provethat the PoA of S1A with monotone subadditive bidders is at least 21 - Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with monotone submodular bidders is at least 1−e1
-
Provethat the PoA of S1A with monotone subadditive bidders is at least 21 - Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
Critique: How about S1A for more general bidders?
S1A for bidders with General (Monotone) Valuations
For a set I of items, a valuation vi:2I→R is monotone, if
for all S⊆T, vi(S)≤vi(T)
S1A for bidders with General (Monotone) Valuations
For a set I of items, a valuation vi:2I→R is monotone, if
for all S⊆T, vi(S)≤vi(T)
Is there a PoA bound for S1A for bidders with general valuations?
Imposssibility Result
With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.
* Even when we consider fixed valuation and MNE.
Imposssibility Result
With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.
The non-deterministic communication complexity of Welfare-Maximization(n) is high
A good PoA bound for simple auctions yields a good communication protocal for Welfare-Maximization(n)
Non-Deterministic Communication Complexity
Fix some function f:X×Y→{0,1}.
Alice gets the input x∈X, Bob gets the input y∈Y, and Charlie gets both x and y. Charlie needs to convince Alice and Bob that f(x,y)=1.
Non-Deterministic Communication Complexity
Fix some function f:X×Y→{0,1}.
Alice gets the input x∈X, Bob gets the input y∈Y, and Charlie gets both x and y. Charlie needs to convince Alice and Bob that f(x,y)=1.
A protocal for Charlie is a function P:{(x,y)∣f(x,y)=1}→{0,1}∗ that "can convince" Alice and Bob.
The complexity of the protocal P is minf(x,y)=1∣P(x,y)∣.
The non-deterministic communication complexity of f is the minimum complexity over all protocal.
Non-Deterministic Communication Complexity
Fix some function f:X×Y→{0,1}.
Alice gets the input x∈X, Bob gets the input y∈Y, and Charlie gets both x and y. Charlie needs to convince Alice and Bob that f(x,y)=1.
A protocal for Charlie is a function P:{(x,y)∣f(x,y)=1}→{0,1}∗ that "can convince" Alice and Bob.
The complexity of the protocal P is minf(x,y)=1∣P(x,y)∣.
The non-deterministic communication complexity of f is the minimum complexity over all protocal.
Example: NDISJ(n)
Non-Deterministic Communication Complexity
Fix some function f:X×Y→{0,1}.
Alice gets the input x∈X, Bob gets the input y∈Y, and Charlie gets both x and y. Charlie needs to convince Alice and Bob that f(x,y)=1.
A protocal for Charlie is a function P:{(x,y)∣f(x,y)=1}→{0,1}∗ that "can convince" Alice and Bob.
The complexity of the protocal P is minf(x,y)=1∣P(x,y)∣.
The non-deterministic communication complexity of f is the minimum complexity over all protocal.
There can be more than two parties!
Welfare-Maximization(n)
There are n players and m items. Each player has some general monotone valuation for each subset of items.
The goal is to identify the case:
(1) Every allocation of items (partition of items into T1,T2,...,Tn) has welfare at most 1.
from
(0) There is some allocation with welfare at least n.
Welfare-Maximization(n)
There are n players and m items. Each player has some general monotone valuation for each subset of items.
The goal is to identify the case:
(1) Every allocation of items (partition of items into T1,T2,...,Tn) has welfare at most 1.
from
(0) There is some allocation with welfare at least n.
We will prove that the non-deterministic complexity of this "promise" problem is exp(Ω(m/n2)) by reduction!
Multi-Disjointness
There are n parties, and each party has a string xi∈{0,1}l(viewed as a subset Si⊆[l])
The goal is to identify the case:
(1) For all i,i′, Si∩Si′=∅
from
(0) ∩i=1nSi=∅
Multi-Disjointness
There are n parties, and each party has a string xi∈{0,1}l(viewed as a subset Si⊆[l])
The goal is to identify the case:
(1) For all i,i′, Si∩Si′=∅
from
(0) ∩i=1nSi=∅
Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)
Multi-Disjointness
There are n parties, and each party has a string xi∈{0,1}l(viewed as a subset Si⊆[l])
The goal is to identify the case:
(1) For all i,i′, Si∩Si′=∅
from
(0) ∩i=1nSi=∅
Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)
Questions?
Multi-Disjointness
There are n parties, and each party has a string xi∈{0,1}l(viewed as a subset Si⊆[l])
The goal is to identify the case:
(1) For all i,i′, Si∩Si′=∅
from
(0) ∩i=1nSi=∅
Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)
Reduction, starts!
Intersecting Family of Partitions
We say t partitions P1,P2,...,Pt of m items into n players are an "intersecting family", if for all i=i′,j=j′,Pji∩Pj′i′=∅
Example: m=6,n=3
P1=[(1,2),(3,4),(5,6)]
P2=[(3,6),(1,5),(2,4)]
P3=[(4,5),(2,6),(1,3)]
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Prove by the probabilistic method!
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Prove by the probabilistic method!
We choose t (a number to be determined) partitions randomly and independently:
In each partition, each item is assigned to one player, independently and uniformly at random.
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Prove by the probabilistic method!
We choose t (a number to be determined) partitions randomly and independently.
For some i=i′,j=j′,
Pr[Pji∩Pj′i′=∅]=(1−n21)m≤e−m/n2
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Prove by the probabilistic method!
We choose t (a number to be determined) partitions randomly and independently.
For some i=i′,j=j′,
Pr[Pji∩Pj′i′=∅]=(1−n21)m≤e−m/n2
Union bound:
Pr[∃i=i′,j=j′s.t.Pji∩Pj′i′=∅] ≤t2n2e−m/n2
Lemma
For every m,n≥1, there exists an intersecting family P1,P2,...,Pt with t=exp(Ω(m/n2)).
Prove by the probabilistic method!
Pr[∃i=i′,j=j′s.t.Pji∩Pj′i′=∅] ≤t2n2e−m/n2
Thus for t<n1em/2n2, the probability (of not being an intersecting family) is less than 1, and the lemma is proven.
Multi-Disjointness to Welfare Maximization
Let S1,S2,...,Sn be the input to the Multi-Disjointness problem with t-bit inputs, where t=exp(Ω(m/n2)) is the same value as in the lemma.
The players coordinate an intersecting family P1,P2,...,Pt in advance.
Multi-Disjointness to Welfare Maximization
Let S1,S2,...,Sn be the input to the Multi-Disjointness problem with t-bit inputs, where t=exp(Ω(m/n2)) is the same value as in the lemma.
The players coordinate an intersecting family P1,P2,...,Pt in advance.
When each player i gets the input Si, one forms the following valuation:
vi(T)={10if T⊇Pij for some j∈Si otherwise.
Multi-Disjointness to Welfare Maximization
Let S1,S2,...,Sn be the input to the Multi-Disjointness problem with t-bit inputs, where t=exp(Ω(m/n2)) is the same value as in the lemma.
The players coordinate an intersecting family P1,P2,...,Pt in advance.
When each player i gets the input Si, one forms the following valuation:
vi(T)={10if T⊇Pij for some j∈Si otherwise.
(1) For all i,i′, Si∩Si′=∅ ⟹ the maximum welfare is at most 1
(0) ∩i=1nSi=∅ ⟹ the maximum welfare is at least n
Multi-Disjointness to Welfare Maximization
Let S1,S2,...,Sn be the input to the Multi-Disjointness problem with t-bit inputs, where t=exp(Ω(m/n2)) is the same value as in the lemma.
The players coordinate an intersecting family P1,P2,...,Pt in advance.
When each player i gets the input Si, one forms the following valuation:
vi(T)={10if T⊇Pij for some j∈Si otherwise.
(1) For all i,i′, Si∩Si′=∅ ⟹ the maximum welfare is at most 1
(0) ∩i=1nSi=∅ ⟹ the maximum welfare is at least n
Reduction complete!
Since we have the Ω(l/n) bound for Multi-Disjointness, the non-deterministic communication complexity of Welfare-Maximization(n) is Ω(t/n)=exp(Ω(m/n2)).
Imposssibility Result
With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.
The non-deterministic communication complexity of Welfare-Maximization(n) is exp(Ω(m/n2))
A good PoA bound for simple auctions yields a good communication protocal for Welfare-Maximization(n)
?
Theorem
Fix a class V of possible bidder valuations.
Suppose that, for some α∈(0,1], there is no nondeterministic communication protocal with subexponential (in m) communication for the following problem:
(1) Every allocation has welfare at most α⋅W∗
(0) There is some allocation with welfare at least W∗
Let ϵ≥1/poly(m,n). Then, for every auction with sub-doubly-exponential (in m) strategies per player, the worst-case PoA of ϵ-approximate MNE with bidder valuations in V is at most α.
Theorem
Fix a class V of possible bidder valuations.
Suppose that, for some α∈(0,1], there is no nondeterministic communication protocal with subexponential (in m) communication for the following problem:
(1) Every allocation has welfare at most α⋅W∗
(0) There is some allocation with welfare at least W∗
Let ϵ≥1/poly(m,n). Then, for every auction with sub-doubly-exponential (in m) strategies per player, the worst-case PoA of ϵ-approximate MNE with bidder valuations in V is at most α.
Theorem
Fix a class V of possible bidder valuations.
Suppose that, for some α∈(0,1], there is no nondeterministic communication protocal with subexponential (in m) communication for the following problem:
(1) Every allocation has welfare at most α⋅W∗
(0) There is some allocation with welfare at least W∗
Let ϵ≥1/poly(m,n). Then, for every auction with sub-doubly-exponential (in m) strategies per player, the worst-case PoA of ϵ-approximate MNE with bidder valuations in V is at most α.
Theorem
Fix V to be general monotone valuations.
Suppose that, for some α∈(0,1], there is no nondeterministic communication protocal with subexponential (in m) communication for the following problem:
(1) Every allocation has welfare at most α⋅W∗
(0) There is some allocation with welfare at least W∗
Let ϵ≥1/poly(m,n). Then, for every auction with sub-doubly-exponential (in m) strategies per player, the worst-case PoA of ϵ-approximate MNE with bidder valuations in V is at most α.
Theorem
Fix V to be general monotone valuations.
Suppose that, for some α∈(0,1], there is no nondeterministic communication protocal with subexponential (in m) communication for the following problem:
(1) Every allocation has welfare at most α⋅W∗
(0) There is some allocation with welfare at least W∗
Let ϵ≥1/poly(m,n). Then, for every auction with sub-doubly-exponential (in m) strategies per player, the worst-case PoA of ϵ-approximate MNE with bidder valuations in V is at most α.
With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible!
Lemma
For every ϵ>0 and every game with n players with strategy sets A1,A2,...,An , there exists an ϵ-approximate MNE with description length polynomial in n, log(max∣Ai∣), and 1/ϵ .
Lemma
For every ϵ>0 and every game with n players with strategy sets A1,A2,...,An , there exists an ϵ-approximate MNE with description length polynomial in n, log(max∣Ai∣), and 1/ϵ .
Proof of Theorem:
So if max∣Ai∣ is sub-doubly-exponential, we can write down the ϵ−approximate MNE, and the contribution to the welfare of each player under the ϵ−approximate MNE in subexponential bits.
If PoA >α, this is in fact a proof for "every allocation has welfare at most α⋅W∗"(from "There is some allocation with welfare at least W∗").
Proof of Theorem:
So if max∣Ai∣ is sub-doubly-exponential, we can write down the ϵ−approximate MNE, and the contribution to the welfare of each player under the ϵ−approximate MNE in subexponential bits..
If PoA >α, this is in fact a proof for "every allocation has welfare at most α⋅W∗"(from "There is some allocation with welfare at least W∗").
The players only need to check:
1. The declared ϵ−approximate MNE is "true" for him/her.
2. The declared contribution to the welfare of his/hers is "true"
3. The sum of all declared players' contribution to the welfare is smaller than α⋅W∗
Proof of Theorem:
So if max∣Ai∣ is sub-doubly-exponential, we can write down the ϵ−approximate MNE, and the contribution to the welfare of each player under the ϵ−approximate MNE in subexponential bits..
If PoA >α, this is in fact a proof for "every allocation has welfare at most α⋅W∗"(from "There is some allocation with welfare at least W∗").
The players only need to check:
1. The declared ϵ−approximate MNE is "true" for him/her.
2. The declared contribution to the welfare of his/hers is "true"
3. The sum of all declared players' contribution to the welfare is smaller than α⋅W∗
Why is this a valid proof to the players?
Proof of Theorem:
So if max∣Ai∣ is sub-doubly-exponential, we can write down the ϵ−approximate MNE, and the contribution to the welfare of each player under the ϵ−approximate MNE in subexponential bits..
If PoA >α, this is in fact a proof for "every allocation has welfare at most α⋅W∗"(from "There is some allocation with welfare at least W∗").
The players only need to check:
1. The declared ϵ−approximate MNE is "true" for him/her.
2. The declared contribution to the welfare of his/hers is "true"
3. The sum of all declared players' contribution to the welfare is smaller than α⋅W∗
But we assumed there is no subexponential communication protocal! ⟹PoA≤α
What we've done
- Prove that the PoA of single-item first-price auction is at least 1−e1
- Prove that the PoA of S1A with monotone submodular bidders is at least 1−e1
-
Provethat the PoA of S1A with monotone subadditive bidders is at least 21 - Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
- Prove that no PoA bounds holds for "simple auctions" for general monotone valuations (considering approximate equilibrium).
Open Problems
- Can we prove that no PoA bound holds for general monotone valuations considering exact equilibria?
- What is the PoA of first-price single-item auctions? (Currently known in [1−e1,0.87])
- PoA in reverse auctions?
AGT beyond PoA of Auctions
- Complexity of computing a Nash Equilibrium (PLS, PPAD)
- Algorithmic Mechanism Design (Knapsack auction)
- ...
End!
Kolmogorov Complexity and the Incompressibility Method
Eric Xiao 2024/8/22
What's the next number?
1,2,4,8,16,
What's the next number?
1,2,4,8,16,
7122?
32?
What's the next number?
1,2,4,8,16,
7122?
32?
"less random"?
What is randomness?
What is randomness?
Random objects are harder to describe/compress?
1,2,4,8,16,7122
1,2,4,8,16,32
A sequence a of length 6 where a1=1 and ai+1=2ai.
The sequence 1, 2, 4, 8, 16, 7122.
A sequence a of length 6 where a1=1 and ai+1=2ai but a6=7122.
1,2,4,8,16,7122
1,2,4,8,16,32
k = 1 for i in range(5): print(k) k *= 2
k = 1 for i in range(4): print(k) k *= 2 print(7122)
A = [1,2,4,8,16,7122] for i in range(5): print(A[i])
Kolmogorov Complexity:
How short can a string be compressed?
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
Kolmogorov Complexity
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
A more formal definition
Cf(x)=min{l(p):f(p)=x}
Where f is some (partial) function.
Kolmogorov Complexity
C(x)=min{l(p):p is an encoding/program of x}
Think of x,p as binary strings(or natural numbers).
A more formal definition
Cf(x)=min{l(p):f(p)=x}
Where f is some (partial) function.
How should we choose f? (English, python, Eric Xiao, ...?)
Kolmogorov Complexity
Maybe f should be the "minimal" (additively optimal) function:
∀g,x,Cf(x)≤Cg(x)+cg
Kolmogorov Complexity
Such f doesn't exist:
Let g(i)=xi, where Cf(xi)≥i
Maybe f should be the "minimal" (additively optimal) function:
∀g,x,Cf(x)≤Cg(x)+cg
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Let f0 be a function computed by a universal TM (T0) with input 11..110np (with l(n) ones ) which simulates Tn on input p
Kolmogorov Complexity
Perhaps we should not consider all functions.
Computable functions?
Let f0 be a function computed by a universal TM (T0) with input 11..110np (with l(n) ones ) which simulates Tn on input p
We have Cf0(x)≤Cfn(x)+2l(n)+1
(since f0(11..10np)=fn(p))
f0 is additively optimal!
?
Kolmogorov Complexity
C(x)=Cf0(x)=min{l(p):f0(p)=x}
Kolmogorov Complexity
C(x)=Cf0(x)=min{l(p):f0(p)=x}
We can view f0 as a special compiler.
Is the compiler of C/Java additively optimal?
Conditional Kolmogorov Complexity
C(x∣y)=Cf0(x∣y)=min{l(p):f0(⟨y,p⟩)=x}
C(x)=C(x∣ϵ)
We also have ∀f,x,y
Cf0(x∣y)≤Cf(x∣y)+cf
Upper bounds
C(x)≤l(x)+c
C(x∣y)≤C(x)+c
Uncomputability
C(x) is uncomputable
Uncomputability
C(x) is uncomputable
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
Uncomputability
C(x) is uncomputable
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
Berry's Paradox:"The smallest positive integer not definable in under sixty letters."
Uncomputability
C(x) is uncomputable on any infinite set of points
function GenerateComplexString() for i = 1 to infinity: for each string s of length exactly i if KolmogorovComplexity(s) ≥ 8000000000 return s
(Assume the length of KolmogorovComplexity() is 700000000)
?
Incompressibility Theorem ∞⋆
Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Incompressibility Theorem ∞⋆
Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
When c=0? c=1?
Incompressibility Theorem ∞⋆
Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Because there are only i=0∑logm−c−12i=m2−c−1 strings with length less than logm−c.
Example - Infinite Primes
Assume that there are only k primes. Then for each n∈N, we can write n=i=1∏kpiαi. Thus we can describe every integer in [1,t] by kloglogt bits......
Incompressibility theorem: Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
Example - Infinite Primes
Assume that there are only k primes. Then for each n∈N, we can write n=i=1∏kpiαi. Thus we can describe every integer in [1,t] by kloglogt bits......
Incompressibility theorem: Let c∈N and A be a set of size m,
then there are at least m(1−2−c)+1 elements x∈A such that
C(x∣y)≥logm−c, for any fixed y.
π(n)∈Ω(loglognlogn)
Example - TM complexity
Theorem: Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Example - TM complexity
Theorem: Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
eg. RADAR
Example - TM complexity
Theorem: Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Assume a TM T:
- each state can be described by ℓ bits
- always terminates with the reading head at the rightmost position of the input
- computes PALINDROME
Example - TM complexity
Theorem: Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Assume a TM T:
- each state can be described by ℓ bits
- always terminates with the reading head at the rightmost position of the input
- computes PALINDROME
Define the crossing sequence of a grid g: the sequence of states of the TM T after the reading head crosses the right boundary of g
Example - TM complexity
By the incompressibility theorem, we can find length-n string x with C(x∣T,n)≥n.
*If the runtime of the TM T is less than ℓn2 for all strings of length 4n, then we can describe every string of length n with 2n+O(logn) bits.
Theorem: Checking if a string is palindrome needs Ω(n2) time for any one-tape Turing Machine.
Example - TM complexity
Consider running T on x02nxˉ, if the runtime is less than ℓn2, then there exists a "middle" grid g0 with crossing sequence shorter than 2ℓn.
In fact, given T,n, we can recover x by the position of g0(O(logn) bits) and the crossing sequence of g0(2n bits).
Thus C(x∣T,n)≤2n+O(logn)
*If the runtime of the TM T is less than ℓn2 for all string of length 4n, then we can describe every string of length n with 2n+O(logn) bits.
Incompressibility Method
If the statement doesn't hold, then there will be too many effectively compressible objects.
?
More Examples: Combinatorial Properties
Some Useful Bounds
en−1nn≤n!≤en−1nn+1
(kn)k≤Ckn≤k!nk<(kne)k
Transitive Tournaments
Tournament: complete directed graph
Transitive tournament: acyclic tournament
v(n): Largest integer such that every tournament on n nodes contains a transitive subtournament on v(n) nodes.


1
2
4
3
Transitive Tournament
Tournament: complete directed graph
Transitive tournament: acyclic tournament
v(n): Largest integer such that every tournament on n nodes contains a transitive subtournament on v(n) nodes.
⌊logn⌋≤ v(n)≤1+⌊2logn⌋
Counting
v(n)≤1+⌊2logn⌋
v:=2+⌊2logn⌋
A: subset of n vertices of size v
σ: permutation on [v]
ΓA,σ: the set of n- vertex tournaments where A is a subtournament of it with "order" σ
We count the number d(Γ′) of n- vertex tournaments with a transitive subtournament of size v
d(Γ′)≤∑A∑σd(ΓA,σ)=Cvnv!2C2n−C2v<2C2n
(Cvnv!<nv≤2C2v)
Probabilistic Method
v(n)≤1+⌊2logn⌋
v:=2+⌊2logn⌋
A: subset of n vertices of size v
σ: permutation on [v]
Let T be the random variable uniformly distributed over all n-vertex tournaments.
We calculate the probability P(T∈Γ′).
P(T∈Γ′)≤∑A∑σP(T∈ΓA,σ)=Cvnv!2−C2v<1
(Cvnv!<nv≤2C2v)
Incompressibility Method
v(n)≤1+⌊2logn⌋
If v(n) is too large, then we can effectively compress every n-vertex tournament.
However, by the incompressibility method, there is some tournament T′, where C(T′∣n,p)≥n(n−1)/2.
(p is some decoding program)
For every tournament T of size n, we can describe it by v(n)⌊logn⌋+n(n−1)/2−v(n)(v(n)−1)/2 bits.
Thus v(n)⌊logn⌋−v(n)(v(n)−1)/2≥0, and
v(n)≤2⌊logn⌋+1.
?
Last Example:
Regular Languages
Regular Language
a language accepted by some DFA

a language decided by some constant-memory one-pass algorithm.
⟺
Regular Language
a language accepted by some DFA

a language decided by some constant-memory one-pass algorithm.
⟺
e.g. L={0∗1}={1,01,001,0001,...}
How to prove that a language is regular?
How to prove that a language is regular?
Construct a DFA/constant-memory one-pass algorithm for it.
How to prove that a language is regular?
Construct a DFA/constant-memory one-pass algorithm for it.
How to prove that a language is non-regular?
Pumping Lemma

Pumping Lemma

Might not be easy to use.
Not characterizing.
Example: Another possible way
Is L={0k1k∣k≥1} a regular language?
Example: Another possible way
Is L={0k1k∣k≥1} a regular language?
Suppose L is accepted by some DFA A. Then we can recover k from A and q, the state of A after processing 0k. Thus for all k≥1, C(k)=O(1).
KC-Regularity
Suppose L is regular, and Lx:={y∣xy∈L}. If y is the n-th string in Lx, then C(y)≤C(n)+O(1).
KC-Regularity
Suppose L is regular, and Lx:={y∣xy∈L}. If y is the n-th string in Lx, then C(y)≤C(n)+O(1).
*What does "the n-th string" mean?
More Examples
Is L:={1p∣p is prime} a regular language?
Is L:={xxˉ} a regular language?
Regular KC-Characterization
In fact, we have
L is regular
For all x,n, if y is the n-th string in Lx, then C(y)≤C(n)+O(1).
⇔
Thus, (in theory), we can also prove regularity!
Takeaways?
1. What is Kolmogorov complexity?
The Kolmogorov complexuty of a string (or number or sequence) is the minimum number of bits it can be compressed to.
2. What is the incompressibility method?
If an object satisfies some property, it can be efficiently compressed, but not everything can be efficiently compressed.
Some Questions
1. Alice and Bob are playing the number guessing game. Alice chooses some x∈[n], and Bob can ask her yes/no questions. If Alice can lie once, what is the minimum number c where Bob can always guess the right answer after asking c questions?
4. Why is computability important?
5. Ask a question!
3. Prove any (nontrivial?) statement you like via the incompressibility method.
2. What are the differences between counting, the probabilistic method, and the incompressibility method?
References
[1] Wikipedia contributors. (2023, March 5). Incompressibility method. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Incompressibility_method&oldid=1143081380
[2] Wikipedia contributors. (2023, August 22). Kolmogorov complexity. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Kolmogorov_complexity&oldid=1171617849
[3] Ming Li , Paul Vitányi. An Introduction to Kolmogorov Complexity and Its Applications, 2008.
A Little Bit of Computational Complexity
By Zi-Hong Xiao
A Little Bit of Computational Complexity
APCS Camp 資料結構
- 886