A Little Bit of Computational Complexity

蕭梓宏

Some interesting questions

  1. P=?NP{\rm P \stackrel{?}{=} NP}
  2. Can we generate mathematical proofs automatically?
  3. Can we verify a proof looking only at very few locations of the proof?
  4. Does random speed up computation?
  5. Is approximation easier?
  6. How powerful is quantum computation?

Computation and Turing Machine

Turing Machine

A (k-tape) turing machine consists of:

1. k tapes that contain symbols(one of them is the input tape)

2. k read-write heads on each tape(the one on the input head is read-only)

3. a register that keeps the state of the TM

4. a finite table of instructions that given the current state and the symbols reading: write a symbol under each read/write head, move the read-write heads , and change the state
 

At first, the TM is in the starting state, and the input tape contains the input. Then, the TM moves according to the table of instructions until it reaches the halting state.

Turing Machine

Formally, a (k-tape) turing machine MM is defined by a tuple(Γ,Q,δ)(\Gamma,Q,\delta):

Γ\Gamma is a finite set,  containing symbols that MM's tapes can contain

QQ is a finite set, containing the states that MM can be in

δ\delta is a function from Q×ΓkQ \times \Gamma^k to Q×Γk1×{L,S,R}kQ \times \Gamma^{k-1} \times \{L,S,R\}^k, and is called the "transition function"

We normally assume that Γ={BLANK,START,0,1}\Gamma=\{\text{BLANK},\text{START},0,1\}, and QQ contains qstartq_{start} and qhaltq_{halt}

Computing a function/Running TIme

Given a function f:{0,1}{0,1}f : \{0,1\}^* \to \{0,1\}^* and another function T:NNT: \mathbb N \to \mathbb N,

we say a TM MM computes ff, if for every x{0,1}x \in \{0,1\}^*, whenever MM is initialized to the start configuration on input xx, then it halts with f(x)f(x) written on its output tape.

 

We say  MM computes ff in T(n)T(n)-time, if its computation on every input xx requires at most T(x)T(|x|) steps.

Some useful facts about TM

  1. There exists an equivalent TM for any (C,java,...) program.(Operations in the machine language can be simulated by TMs) 
  2. If ff is computable on a k-tape TM in T(n)T(n)- time, then ff is computable in time 5kT(n)25kT(n)^2 on a single-tape TM whose head movements don't depend on the input and only depends on the input length.
  3. Each string represents a TM, and each TM can be represented by infinitely many strings

Universal TM

There exsists TM UU such that for any x,α{0,1}x,\alpha \in \{0,1\}^* ,U(x,α)=Mα(x)U(x,\alpha) = M_\alpha(x), where MαM_\alpha is the TM represented by α\alpha.

Moreover, if Mα(x)M_\alpha(x) halts within TT steps then U(x,α)U(x,\alpha) halts within CTlogTCT\log T steps, where CC is independent of x|x|

CT2CT^2 instead of CTlogTCT\log T:

Universal TM with Time Bound

There exsists TM UU such that for any x,α{0,1}x,\alpha \in \{0,1\}^* , TNT \in \mathbb N,U(x,α,T)=Mα(x)U(x,\alpha,T) = M_\alpha(x) if Mα(x)M_\alpha(x) halts in TT steps

Uncomputability/Undecidability?

演算法,問題不可勝數?

Uncomputable function

Define the function UC:{0,1}{0,1}{\rm UC}: \{0,1\}^* \to \{0,1\}:

if Mα(α)=1M_\alpha(\alpha) = 1,UC(α)=0{\rm UC}(\alpha) = 0;otherwise, UC(α)=1{\rm UC}(\alpha) = 1

No turing machine can compute UC\rm UC!!!

Little fact: check if integer coefficient polynomial equations have integer solution is also uncomputable

Halting problem

HALT:{0,1}{0,1}{\rm HALT}: \{0,1\}^* \to \{0,1\}:

HALT(α,x)=1{\rm HALT}(\langle \alpha,x \rangle) = 1, if Mα(x)M_\alpha(x) halts;

HALT(α,x)=0{\rm HALT}(\langle \alpha,x\rangle) = 0, otherwise

HALT{\rm HALT} is uncomputable. WHY?

HALT:{0,1}{0,1}{\rm HALT}: \{0,1\}^* \to \{0,1\}:

HALT(α,x)=1{\rm HALT}(\langle \alpha,x \rangle) = 1, if Mα(x)M_\alpha(x) halts;

HALT(α)=0{\rm HALT}(\alpha) = 0, otherwise

If we have MhaltM_{\rm halt}, then we can construct MUCM_{\rm UC}:

1. Run Mhalt(α,α)M_{\rm halt}(\alpha,\alpha) to see if Mα(α)M_\alpha(\alpha) halts

2. If it does, run U(α,α)U(\alpha,\alpha) to get its output

3. Ouput 00 if Mα(α)M_\alpha(\alpha) halts and outputs 11; output 11 otherwise

We can REDUCE(\infty \star) UC{\rm UC} to HALT{\rm HALT}!

UC:{0,1}{0,1}{\rm UC}: \{0,1\}^* \to \{0,1\}:

UC(α)=0{\rm UC}(\alpha) = 0, if Mα(α)=1M_\alpha(\alpha) = 1;

UC(α)=1{\rm UC}(\alpha) = 1, otherwise

Uncomputability and

Gödel's Imcompleteness Theorem?

Gödel's Imcompleteness Theorem

First Incompleteness Theorem: "任何可以表達算術且一致的形式系統一定是不完備的。"

一致(consistent): There is no statement ϕ\phi such that both ϕ\phi and ¬ϕ\neg \phi can be proven

完備(complete): For every statement ϕ\phi, either ϕ\phi or ¬ϕ\neg \phi can be proven

Gödel's Imcompleteness Theorem

There are true statements that cannot be proven!

(in certain axiomatic systems, like ZFC)

For example, the consistency of ZFC cannot be proven in ZFC.

Gödel's imcompleteness theorem motivated Church and Turing into their work on uncomputability.

The other way?

From Halting Problem to
Incompleteness Thorem

Weaker version of First Incompleteness Theorem: "任何可以表達算術且可靠的形式系統一定是不完備的。"

可靠(sound): If a statement ϕ\phi is not true, it cannot be proven(which implies consistent)

完備(complete): For every statement ϕ\phi, either ϕ\phi or ¬ϕ\neg \phi can be proven

From Halting Problem to
Incompleteness Thorem

If a axiomatic system(that is powerful enough) is both sound and complete , we can devise an algorithm(or, a TM) that can compute HALT\rm HALT!

From Halting Problem to
Incompleteness Thorem

step 0: Read the input α\alpha and xx

step 1: Express the statement "MαM_\alpha halts on xx" by ϕ\phi

step 2: Enumerate all strings of finite length, and check if it is a proof for ϕ\phi or ¬ϕ\neg \phi(proof checking is not hard!)

step 3: By completeness, step 2 can find the proof, and by soundness, the proven result is true. Thus we can output 11 if ϕ\phi is true and output 00 otherwise!

Why Turing Machine?

The Church-Turing(CT) Thesis: Every effectively calculable function is a computable function

λ\lambda-calculus

general recursive function

Markov alogorithm

pointer machine

combinatory logic

computationally equivalent to TMs

P

Language/Decision Problem

A language LL is a subset of

{0,1}\{0,1\}^*(the set of all (bit)strings)

A decision problem for LL is:

"Given xx: output 1 if xLx \in L, else output 0"

*We can turn many objects into strings, like integers, pairs of integers, matrix, graph,...,and turing machines!

Language/Decision Problem

Examples of languages:

PAL:{s:s is a palindrome}\{s:s \text{ is a palindrome}\}

PRIME:{n:n is a prime number}\{n: n \text{ is a prime number}\}

MAXINDSET:{G,k: the size of the largest independent set  in G is k}\{\langle G,k\rangle:\text{ the size of the largest independent set } \text{ in } G \text{ is } k\}

Complexity Class

A complexity class is a set of languages(boolean functions) that can be decided(computed) within given resource bounds.

DTIME and P

DTIME(T(n)){\rm DTIME}(T(n)): a language LL is in DTIME(T(n)){\rm DTIME}(T(n)) if there is a TM that decides LL in O(T(n))O(T(n)) time

P: P=c1DTIME(nc)P = \cup_{c \geq 1} {\rm DTIME}(n^c)

We say that problems in  P are efficiently solvable

Examples?

NP

NP

a language LL is in NP\rm NP :

there exists a polynomial pp, a polynomial-time TM MM, such that x{0,1},xLu{0,1}p(x),s.t.M(u,x)=1\forall x \in \{0,1\}^*, \\ x\in L \Leftrightarrow \exists u \in \{0,1\}^{|p(|x|)|},{\rm s.t.} M(u,x)=1

"A 'yes' instance can be verified in polynomial time with a certificate."

NP

Examples of languages in NP:

PAL:{s:s is a palindrome}\{s:s \text{ is a palindrome}\}

COMPOSITE:{n:n is a composite number}\{n: n \text{ is a composite number}\}

SUBSETSUM:{A1,A2,...,An,T: there is a subset of A that sum to T}\{\langle A_1,A_2,...,A_n,T\rangle:\text{ there is a subset of } A \text{ that sum to } T \}

Are they in NP?

INDSET:{G,k: there is a independent set of size k in G}\{\langle G,k\rangle:\text{ there is a independent set of size } k \text{ in } G \}

MAXINDSET:{G,k: the size of the largest independent set  in G is k}\{\langle G,k\rangle:\text{ the size of the largest independent set } \text{ in } G \text{ is } k\}

GI:{G,G:G is isomorphic to G}\{\langle G,G'\rangle: G \text{ is isomorphic to } G' \}
GNI:{G,G:G is not isomorphic to G}\{\langle G,G'\rangle: G \text{ is not isomorphic to } G' \}

?

?

Non-Deterministic TM

An NDTM MM is almost the same as a TM, but it has a special state qacceptq_{\rm accept} and two transition functions δ0\delta_0 and δ1\delta_1. In each step of computation, MM arbitrarily choose one of them to apply.

For x{0,1}x\in\{0,1\}^*, we say that M(x)=1M(x) = 1 if there exists some sequence of these choices that makes MM reach qacceptq_{\rm accept}.(Otherwise, M(x)M(x) = 0)

We say that MM runs in T(n)T(n) time, if for all x{0,1}x\in \{0,1\}^* and every sequence of choices, MM reaches either qhaltq_{\rm halt} or  qacceptq_{\rm accept} within T(x)T(|x|) steps.

NTIME

NTIME(T(n)){\rm NTIME}(T(n)): a language LL is in NTIME(T(n)){\rm NTIME}(T(n)) if there is a O(T(n))O(T(n))  NDTM MM, such that xLM(x)=1x \in L \Leftrightarrow M(x) = 1

Theorem 1:

NP=c1NTIME(nc){\rm NP} = \cup_{c \geq 1} {\rm NTIME}(n^c)

Idea:

The choices that make the NDTM reach qacceptq_{\rm accept} can be used as the certificate(and vice versa).

Polynomial-time Reduction

A language LL is polynomial-time reducible to another language LL'(LpLL \leq_p L'), if there exists a polynomial-time computable function f:{0,1}{0,1}f:\{0,1\}^* \to \{0,1\}^* such that xLf(x)Lx \in L \Leftrightarrow f(x) \in L'

LpL,LpLLpLL \leq_p L', L' \leq_p L'' \Rightarrow L \leq_p L''

example: we can reduce 3-coloring to 4-coloring.

NP hard and NP complete

NP-hard: LL' is in NP-hard, if LpLL \leq_p L' for every LL in NP

NP-complete: LL' is NP-complete if it is in NP and NP-hard 

CNF(CNF formula)

a CNF instance with 5 variables and 4 clauses: 

ϕ(a1,a2,a3,a4,a5)=(a1¬a2a3a4)(¬a1a2a3a5)(¬a1¬a4a5)(a2¬a3¬a5)\phi(a_1,a_2,a_3,a_4,a_5) = (a_1\lor \neg a_2 \lor a_3 \lor a_4) \land (\neg a_1 \lor a_2 \lor a_3 \lor a_5) \land (\neg a_1 \lor \neg a_4 \lor a_5) \land (a_2 \lor \neg a_3 \lor \neg a_5)

a 3CNF instance with 5 variables and 4 clauses: 

ϕ(a1,a2,a3,a4,a5)=(a1¬a2a3)(¬a1a2a3)(¬a1¬a4a5)(a2¬a3¬a5)\phi(a_1,a_2,a_3,a_4,a_5) = (a_1\lor \neg a_2 \lor a_3) \land (\neg a_1 \lor a_2 \lor a_3 ) \land (\neg a_1 \lor \neg a_4 \lor a_5) \land (a_2 \lor \neg a_3 \lor \neg a_5)

A CNF is satisfiable if there exists some assignment of the variables such that ϕ\phi is true.

clause

literal

SAT and 3SAT

SAT:{ϕ:ϕ is a satisfiable CNF}\{\phi: \phi \text{ is a satisfiable CNF}\}

3SAT:{ϕ:ϕ is a satisfiable 3CNF}\{\phi: \phi \text{ is a satisfiable 3CNF}\}

Cook-Levin Theorem:

SAT and 3SAT is NP-complete.

We can see that SAT and 3SAT is in NP.

Our goal is to

1. reduce every LNPL \in {\rm NP} to SAT,

2. reduce SAT to 3SAT

useful tool

Every boolean function f:{0,1}l{0,1}f:\{0,1\}^l \to \{0,1\} can be expressed as a CNF with 2l2^l clauses of ll literals

ϕ(x)=z:f(z)=0(xz)\phi(x) = \bigwedge\limits_{z:f(z) = 0} (x \neq z)

x_1 x_2 x_3 f(x)
0 0 0 1
0 0 1 0
0 1 0 1
0 1 1 0
1 0 0 0
1 0 1 1
1 1 0 1
1 1 1 1

ϕ(x)=(x1x2¬x3)(x1¬x2¬x3)(¬x1x2x3)\phi(x) = (x_1 \lor x_2 \lor \neg x_3) \land\\ (x_1 \lor \neg x_2 \lor \neg x_3)\land\\ (\neg x_1 \lor x_2 \lor x_3)

CL Theorem Proof

We want to reduce a language LL in NP to SAT(in polynomial time).

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

Since LL is in NP, there is a polynomial time TM MM such that

xLM(x,u)=1x \in L \Leftrightarrow M(x,u)=1 for some u{0,1}p(x)u \in \{0,1\}^{p(|x|)}

CL Theorem Proof

We can assume that:

MM has only two tapes (input + work / output) and it is oblivious(so the location of its heads in the ii-th step only depends on the input length and ii)

CL Theorem Proof

Snapshot:

The snapshot of the ii-th step of MM's computation on input y=xuy = x \circ u is a triple a,b,qΓ×Γ×Q\langle a,b,q \rangle \in \Gamma \times \Gamma \times Q, representing the symbols on the two tapes and the state that MM is in.

Encode the snapshot of the ii-th step as zi{0,1}cz_i \in \{0,1\}^c, where cc is only dependent of Γ|\Gamma| and Q|Q|.

If someone claims that M(x,u)=1M(x,u) = 1 and provides you the snapshot of the computation, how can you tell whether the snapshots present a valid computution by MM

CL Theorem Proof

ziz_i depends on: the state in step i1i-1 and the symbols in the currenct cells of the two tapes

There is a function F:{0,1}2c+1{0,1}cF:\{0,1\}^{2c+1} \to \{0,1\}^c, such that

zi=F(zi1,zprev(i),yinputpos(i))z_i=F(z_{i-1},z_{{\rm prev}(i)},y_{{\rm inputpos}(i)})

FF only depends on MM; prev(i){\rm prev}(i) and inputpos(i){\rm inputpos}(i) only depend on MM and y|y|, and can be known by simulating M(0y)M(0^{|y|}).

CL Theorem Proof

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

M(x,u)=1 for some u{0,1}p(x)M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)} \Leftrightarrow

there exist y{0,1}x+p(x)y \in \{0,1\}^{|x|+p(|x|)} and z{0,1}c(T(n)+1)z \in \{0,1\}^{c(T(n)+1)}, such that

1. The first nn bits of yy are equal to xx

2. z0z_0 encodes the initial snapshot of MM

3. For every i[1,T(n)],zi=F(zi1,zprev(i),yinputpos(i))i\in [1,T(n)],z_i = F(z_{i-1},z_{{\rm prev}(i)},y_{{\rm inputpos}(i)})

4. zT(n)z_{T(n)} encodes a snapshot in which MM halts and outputs 11.

CL Theorem Proof

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

M(x,u)=1 for some u{0,1}p(x)M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)} \Leftrightarrow

there exist y{0,1}x+p(x)y \in \{0,1\}^{|x|+p(|x|)} and z{0,1}c(T(n)+1)z \in \{0,1\}^{c(T(n)+1)}, such that

1. "usefull tool": nn clauses of size 44

2. z0z_0 encodes the initial snapshot of MM

3. For every i[1,T(n)],zi=F(zi1,zprev(i),yinputpos(i))i\in [1,T(n)],z_i = F(z_{i-1},z_{{\rm prev}(i)},y_{{\rm inputpos}(i)})

4. zT(n)z_{T(n)} encodes a snapshot in which MM halts and outputs 11.

CL Theorem Proof

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

M(x,u)=1 for some u{0,1}p(x)M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)} \Leftrightarrow

there exist y{0,1}x+p(x)y \in \{0,1\}^{|x|+p(|x|)} and z{0,1}c(T(n)+1)z \in \{0,1\}^{c(T(n)+1)}, such that

1. "usefull tool": nn clauses of size 44

2. "useful tool": 2c2^c clauses of size cc

3. For every i[1,T(n)],zi=F(zi1,zprev(i),yinputpos(i))i\in [1,T(n)],z_i = F(z_{i-1},z_{{\rm prev}(i)},y_{{\rm inputpos}(i)})

4. zT(n)z_{T(n)} encodes a snapshot in which MM halts and outputs 11.

CL Theorem Proof

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

M(x,u)=1 for some u{0,1}p(x)M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)} \Leftrightarrow

there exist y{0,1}x+p(x)y \in \{0,1\}^{|x|+p(|x|)} and z{0,1}c(T(n)+1)z \in \{0,1\}^{c(T(n)+1)}, such that

1. "usefull tool": nn clauses of size 44

2. "useful tool": 2c2^c clauses of size cc

3. "useful tool": T(n)23c+1T(n) 2^{3c+1} clauses of size 3c+13c+1

4. zT(n)z_{T(n)} encodes a snapshot in which MM halts and outputs 11.

CL Theorem Proof

Given x{0,1}x \in \{0,1\}^*, we want to construct ϕx\phi_x such that

ϕxSATM(x,u)=1 for some u{0,1}p(x)\phi_x \in {\rm SAT} \Leftrightarrow M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)}

M(x,u)=1 for some u{0,1}p(x)M(x,u) = 1 \text{ for some } u \in \{0,1\}^{p(|x|)} \Leftrightarrow

there exist y{0,1}x+p(x)y \in \{0,1\}^{|x|+p(|x|)} and z{0,1}c(T(n)+1)z \in \{0,1\}^{c(T(n)+1)}, such that

1. "usefull tool": nn clauses of size 44

2. "useful tool": 2c2^c clauses of size cc

3. "useful tool": T(n)23c+1T(n) 2^{3c+1} clauses of size 3c+13c+1

4. "useful tool": 2c2^c clauses of size cc

Just let ϕx\phi_x be the AND of the above conditions!

NP completeness of 3SAT

Now we know SAT is NP-complete, so we can show that 3SAT is NP-complete by reducing SAT to 3SAT in polynomial time.

(a1¬a2a3a4¬a5)(a1¬a2a3b1)(a4¬a5¬b1)(a_1\lor \neg a_2 \lor a_3 \lor a_4 \lor \neg a_5) \\ \Leftrightarrow (a_1\lor \neg a_2\lor a_3 \lor b_1) \land (a_4 \lor \neg a_5 \lor \neg b_1)

(a1¬a2b2)(a3b1¬b2)(a4¬a5¬b1)\Leftrightarrow (a_1\lor \neg a_2\lor b_2) \land (a_3 \lor b_1 \lor \neg b_2) \land (a_4 \lor \neg a_5 \lor \neg b_1)

We can easily turn a clause of kk variables into AND of k2k-2 clauses of 33 variables!

Reduction 大賽

INDSET: Decide if there is a independent set of size kk on GG

QUADEQ: Decide if the set of boolean quadratic equations (i,j[n]ai,juiuj=b\sum_{i,j \in [n]} a_{i,j}u_i u_j = b) is satisfiable(addition is modulo 2)

dHAMPATH: Decide if the directed graph GG has a hamiltonian path

Exactly One 3SAT: Decide if there is an assigment such that in each clause of the CNF, exactly one of the literals is TRUE

TSP:  Decide if there is a closed circuit that visits every node exactly once with total length at most kk on a weighted complete graph GG

HAMPATH: Decide if the undirected graph GG has a hamiltonian path

SUBSETSUM: Decide if a subset of the sequence AA sum up to TT

coNP(順便講)

1. coNP = {L:LˉNP}\{L : \bar L \in {\rm NP}\}

2. We say a language LL is in coNP if there are a polynomial time TM MM and a polynomial pp such that

xLu{0,1}p(x),M(x,u)=1x \in L \Leftrightarrow \forall u \in \{0,1\}^{p(|x|)}, M(x,u) = 1

Why are the two defenitions equivalent?

example:GNI,tautology

Hardness of Approximation

MAX-3SAT

For a 3CNF ϕ\phi, we denote val(ϕ){\rm val}(\phi) as the maximum fraction of clauses that can be satisfied by any assigment.

If ϕ\phi is satisfiable, then val(ϕ)=1{\rm val}(\phi) = 1.

ρ\rho-approximation

For a CNF ϕ\phi, can we find an assignment that satisfies ρval(ϕ)\rho \cdot{\rm val}(\phi) of ϕs\phi's clauses in polynomial time?

For a graph GG whose maximum independent set has size kk, can we find an independent set of size ρk\rho \cdot k in polynomial time?

For a graph GG whose minimum vertex cover has size kk, can we find an vertex cover of size kρ\frac{k}{\rho} in polynomial time?

Can ρ\rho be arbitrarily close to 11?

PCP Theorem

PCP verifier?

Example: GNIPCP(poly(n),1){\rm GNI} \in {\rm PCP}({\rm poly}(n),1)(GNI has a PCP verifier that uses poly(n) random coins and O(1)O(1) queries to the proof)

Verifier(VV): By tosing random coins and making queries to the proof, the verifier should accept if G1≢G2G_1\not \equiv G_2.

Proof(π\pi): The proof is a string, and is expected that for each nn-vertex labeled graph HH, π[H]=0\pi[H] = 0 if HG0H \equiv G_0 and π[H]=1\pi[H] = 1 otherwise.

VV flips a coin to get b{0,1}b \in \{0,1\} and flips O(nlogn)O(n \log n)  coins to get a random permutation PP of length nn.

VV applies PP to the vertices of GbG_b to obtain HH, and accept iff π[H]=b\pi[H] = b.

Efficient: runs in polynomial time and uses limited random coins and queries

Complete: always accepts if xLx \in L

Sound: accepts with probability at most 12\frac{1}{2} if x∉Lx \not \in L

香草口味PCP theorem

NP=PCP(logn,1){\rm NP}={\rm PCP}(\log n,1)

Each language in NP has a PCP verifier that uses O(logn)O(\log n) random coins and makes O(1)O(1) queries to the proof.

We can then probabilisticlly check a mathemaical proof by only examining a constant bit of the proof!

巧克力口味PCP theorem

There exists ρ<1\rho < 1, such that for every LNPL \in \rm NP there is a polynomial function ff mapping strings to 3CNF such that :

xLval(f(x))=1x \in L \Rightarrow {\rm val}(f(x)) = 1

x∉Lval(f(x))<ρx \not\in L \Rightarrow {\rm val}(f(x)) < \rho

Thus, there exists some ρ<1\rho < 1 such that if there is a ρ\rho-approximation algortihm for MAX-3SAT, then P=NP.

Why are they equivalent?

巧克力:There exists ρ<1\rho < 1, such that for every LNPL \in \rm NP there is a polynomial function ff mapping strings to 3CNF such that :

xLval(f(x))=1x \in L \Rightarrow {\rm val}(f(x)) = 1

x∉Lval(f(x))<ρx \not\in L \Rightarrow {\rm val}(f(x)) < \rho

香草:Each language in NP has a PCP verifier VV that uses O(logn)O(\log n) random coins and makes O(1)O(1) queries to the proof.

香草到巧克力:Define Vx,r{0,1}V_{x,r} \in \{0,1\} as the output of VV if the random coin tosses are rr(r{0,1}clognr \in \{0,1\}^{c \log n}). Since Vx,rV_{x,r} only depends on qq bits of the proof, we can express Vx,rV_{x,r} by a 3CNF with q2qq2^q clauses(poly(n)q2q{\rm poly}(n) \cdot q2^q clauses in total).

Why are they equivalent?

巧克力:There exists ρ<1\rho < 1, such that for every LNPL \in \rm NP there is a polynomial function ff mapping strings to 3CNF such that :

xLval(f(x))=1x \in L \Rightarrow {\rm val}(f(x)) = 1

x∉Lval(f(x))<ρx \not\in L \Rightarrow {\rm val}(f(x)) < \rho

香草:Each language in NP has a PCP verifier VV that uses O(logn)O(\log n) random coins and makes O(1)O(1) queries to the proof.

巧克力到香草:For an input xx, the verifier computes f(x)f(x) and expect the proof to be a satisfying assigment for f(x)f(x). Then the verifier just randomly choose one clause of f(x)f(x) and look at the corresponding three bits of the proof to check.

Maxmimum Independent Set and Minimum Vertex Cover

By the reduction of 3SAT to INDSET, we can transform a CNF ϕ\phi to a nn-vertex graph whose largest independent set has size val(ϕ)n7{\rm val}(\phi)\frac{n}{7}

By the PCP theorem, if PNP{\rm P} \neq {\rm NP}, there exists ρ<1\rho < 1 such that we cannot decide if val(ϕ)=1{\rm val}(\phi) = 1 or val(ϕ)<ρ{\rm val}(\phi) < \rho.

val(ϕ){\rm val}(\phi)

最大獨立集

最小點覆蓋

1

<ρ<\rho

n7\frac{n}{7}

<ρn7<\rho\frac{n}{7}

nn7n-\frac{n}{7}

>nρn7>n - \rho \frac{n}{7}

nn7nρn7=67ρ\frac{n-\frac{n}{7}}{n-\rho\frac{n}{7}} = \frac{6}{7-\rho}, thus we cannot ρ\rho-approximate the maximum independent set or 67ρ\frac{6}{7-\rho} -approximate the minimum vertex cover.

Approximation of maximum independent set

In fact, we cannot ρ\rho'-approximate the maximum independent set for any ρ\rho'!

GkG^k

GG

G2G^2

IS(Gk)=(IS(G))k{\rm IS}(G^k) = ({\rm IS}(G))^k

Maxmimum Independent Set and Minimum Vertex Cover

By the reduction of 3SAT to INDSET, we can transform a CNF ϕ\phi to a nn-vertex graph whose largest independent set has size val(ϕ)n7{\rm val}(\phi)\frac{n}{7}

val(ϕ){\rm val}(\phi)

IS(G){\rm IS}(G)

1

<ρ<\rho

n7\frac{n}{7}

<ρn7<\rho\frac{n}{7}

(n7)k(\frac{n}{7})^k

IS(Gk){\rm IS}(G^k)

<ρk(n7)k<\rho^k(\frac{n}{7})^k

For any ρ>0\rho' > 0, there exits kk such that ρk<ρ\rho^k < \rho', thus we cannot ρ\rho'-approximate IS in polynomial time if PNP{\rm P} \neq {\rm NP}

Proof of the PCP theorm

沒了啦><

Reference

Sanjeev Arora and Boaz Barak(2009):Computational Complexity: A Modern Approach 

Communication Complexity?

What is communication complexity?

Communication Complexity

For a function f(x,y)f(x,y), where xX,yYx\in \mathcal X, y \in \mathcal Y, we give xx to Alice(A) and yy to Bob(B), and they both need to know f(x,y)f(x,y)

A protocol is an algorithm that decide how A and B communicate. The length of a protocol is the maximum  number of bits A and B need to exchange over all (x,y)(x,y).

The communication complexity of ff is the minimum length of all protocols that "computes" ff.

Examples!

For x,y{0,1}nx,y\in \{0,1\}^n,

EQ(x,y)={1(if x=y)0(if xy) EQ(x,y)=\left\{\begin{matrix}1 & ({\rm if}\ x=y)\\ 0 & ({\rm if}\ x\neq y)\end{matrix}\right.

Let x,yx,y be lists of numbers in [n][n] whose  length is tt.

MD(x,y)MD(x,y) is the median of the combination of the two lists.
* However, we usually assume f(x,y){0,1}f(x,y) \in \{0,1\}.

Formal Definition of Protocols

A protocol π\pi can be specified by a full binary tree, where each non-leaf vertex has an owner and an associated function fv:X(or Y) {0,1}f_v: \mathcal X( {\rm or}\ \mathcal Y)  \to \{0,1\}, and each leaf has an output.

Each input (x,y)(x,y) induce a path from the root to the leaf π(x,y)\pi(x,y). We say π\pi computes a function ff if the output of π(x,y)\pi(x,y) is equal to f(x,y)f(x,y) for all possible inputs.
 

The length of the protocol is the depth of the tree.

Randomized Protocols?

In the randomized case, A and B can both see an infinitely-long sequence RR of unifromly random bits(independent of their inputs), and the protocol can make errors.

We say a protocol π\pi computes ff with error ϵ<12\epsilon < \frac{1}{2}, if for all (x,y)(x,y), PrR[πR(x,y)f(x,y)]ϵ\Pr_{R}[\pi_R(x,y) \neq f(x,y)] \leq \epsilon

The length of π\pi is the maximum number of bits exchanged(over all R,x,yR,x,y). The randomized communication complexity of f,ϵf,\epsilon is the minimum length of all randomized ptotocol that computes ff with error ϵ\epsilon.

Examples!

For x,y{0,1}nx,y\in \{0,1\}^n,

EQ(x,y)={1(if x=y)0(if xy) EQ(x,y)=\left\{\begin{matrix}1 & ({\rm if}\ x=y)\\ 0 & ({\rm if}\ x\neq y)\end{matrix}\right.

Some Other Problems

For x,y{0,1}nx,y\in \{0,1\}^n,

GT(x,y)={1(if xy)0(if x<y) GT(x,y)=\left\{\begin{matrix}1 & ({\rm if}\ x\geq y)\\ 0 & ({\rm if}\ x < y)\end{matrix}\right.

For x,y{0,1}nx,y\in \{0,1\}^n,

DISJ(x,y)={1(if xy=0)0(otherwise) DISJ(x,y)=\left\{\begin{matrix}1 & ({\rm if}\ |x \cap y|=0)\\ 0 & ({\rm otherwise})\end{matrix}\right.

For x,y{0,1}nx,y\in \{0,1\}^n,

IP(x,y)=x,y%2 IP(x,y)=\langle x,y \rangle \% 2

Lower Bounds

Deterministic

Functions and Matrices

For a function f(x,y)f(x,y), we can represent it as a matrix MfM_f.

MEQM_{EQ}(2-bit)

Rectangles

A rectangle is a  subset of the input space X×Y\mathcal X \times \mathcal Y of the form A×BA\times B, where AX,BYA\in \mathcal X, B \in \mathcal Y.

Rectangles and Protocols

After Alice and Bob communicate according to the protocol, a rectangle is induced.

The rectangle must be monochromatic in MfM_f.

Rectangles and Protocols

Thus, if there exists a protocol of length cc for computing ff, then MfM_f can be partitioned into 2c2^c monochromatic rectangles.

If every partition of MfM_f into monochromatic rectangles requires at least tt rectangles, then the deterministic complexity of ff is at least logt\log t.

    \implies

Rectangles and Protocols

Thus, if there exists a protocol of length cc for computing ff, then MfM_f can be partitioned into 2c2^c monochromatic rectangles.

If every cover of MfM_f by monochromatic rectangles requires at least tt rectangles, then the deterministic complexity of ff is at least logt\log t.

(\infty \star)

    \implies

Rectangles and Protocols

Is the converse true?

However, if MfM_f can be partitioned into 2c2^c rectangles, then there is a protocol of length O(c2)O(c^2) that computes ff.

Deterministic Lower Bound of EQ and GT

MEQM_{EQ}

MGTM_{GT}

Fooling Set

A fooling set FF of f(x,y)f(x,y) is a subset of inputs such that

1. ff is constant on FF, and

2. for all (x1,y1),(x2,y2)F(x_1,y_1),(x_2,y_2) \in F, either (x1,y2)(x_1,y_2) or (x2,y1)(x_2,y_1) has the opposite ff-value

The size of the (monochromatic) rectangle cover of MfM_f is at least the size of F|F|.

The deterministic communication complexity of ff is at least logf\log f.

Deterministic Lower Bound of DISJ

For x,y{0,1}nx,y\in \{0,1\}^n,

DISJ(x,y)={1(if xy=0)0(otherwise) DISJ(x,y)=\left\{\begin{matrix}1 & ({\rm if}\ |x \cap y|=0)\\ 0 & ({\rm otherwise})\end{matrix}\right.

F={(x,xˉ)x{0,1}n}F = \{(x,\bar x)|x \in \{0,1\}^n\} is a fooling set.

The Deterministic Communication Complexity  of DISJ is n+1.

Rank

The deterministic communication complexity of ff is at least log(rank(Mf)+1)\log({\rm rank}(M_f)+1)

rank(Mf){\rm rank}(M_f) is the smallest number rr such that MfM_f can be expressed as the sum of rr matrices of rank 1

Protocol of length cc implies a partition of 2c2^c rectangles.

Thus,  rank(Mf)2c1{\rm rank}(M_f) \leq 2^c-1.

Lower Bounds from Rank

MEQM_{EQ}

MGTM_{GT}

Some Lemmas

rank(MM)=rank(M)rank(M){\rm rank}(M\otimes M') = {\rm rank}(M) \cdot {\rm rank}(M')

rank(A)rank(B)rank(A+B)rank(A)+rank(B){\rm rank}(A) - {\rm rank}(B) \leq {\rm rank}(A+B) \leq {\rm rank}(A) + {\rm rank}(B)

Lower Bounds from Rank

MDISJM_{DISJ}

MIPM_{IP}

Lower Bounds from Rank

MDISJ2M_{DISJ_2}

P=J2MIPP= J-2M_{IP}

Lower Bounds from Rank

MDISJn=[1110]MDISJn1M_{DISJ_n} = \begin{bmatrix} 1&1 \\ 1&0\end{bmatrix} \otimes M_{DISJ_{n-1}}

Pn=[1111]Pn1P_{n} = \begin{bmatrix} 1&1 \\ 1&-1\end{bmatrix} \otimes P_{n-1}

rank(MIPn)rank(Pn)rank(Jn){\rm rank}(M_{IP_n}) \geq {\rm rank} (P_n) - {\rm rank} (J_n)

Randomized

Public VS Private Coin

private-coin protocol of length cc     \implies public-coin protocol of length cc

public-coin protocol of length cc     \implies private-coin protocol of length O(c+logn)O(c+\log n)

(Newman's Theorem)

Randomized Protocol

=

Distribution of Deterministic Protocols

Distributional Complexity

von Neumann's minimax principle: For a n×mn\times m matrix MM

maxxRn,x1=1,xi>0minyRm,y1=1,yi>0xTMy=minymaxxxTMy\max\limits_{x\in \mathbb R^n, ||x||_1 =1, x_i > 0} \min\limits_{y\in \mathbb R^m, ||y||_1 =1, y_i > 0} x^T M y = \min_y \max_x x^T M y

The randomized complexity of f,ϵf,\epsilon is at least cc

\Leftrightarrow

There is a distribution DD of inputs (x,y)(x,y) such that for all deterministic protocol π\pi with error ϵ\epsilon, the length of π\pi on DD is at least cc

\infty \star Yao's(姚期智) Principle

Discrepancy

For a set SX×YS\subseteq \mathcal X \times \mathcal Y, and a distribution μ\mu  over the inputs, we say the discrepancy of f(x,y)f(x,y) w.r.t.  SS and μ\mu is 

Eμ[χS(x,y)(1)f(x,y)]|\mathbb E_\mu [\chi_S(x,y)(-1)^{f(x,y)}]|

low discrepancy for every rectangle     \implies high communication complexity?

For a distribution μ\mu, if the discrepancy of ff w.r.t. every rectangle is at most γ\gamma, then the length of any ptotocol computing ff with error ϵ\epsilon is at least log(12ϵγ)\log(\frac{1-2\epsilon}{\gamma}) when inputs are drawn from μ\mu.

Randomized Lower Bound of IP

For every rectangle RX×YR \subseteq \mathcal X \times \mathcal Y,

EU[χR(x,y)(1)IP(x,y)]2n|\mathbb E_U[\chi_{R}(x,y) (-1)^{IP(x,y)}]| \leq \sqrt{2^{-n}}

The randomized communication complexity of IP is at least n2log(112ϵ)\frac{n}{2}-\log(\frac{1}{1-2\epsilon}) (Ω(n)\Omega(n))

Randomized Lower Bound of DISJ

Ω(n)\Omega(n)

Ω(n)\Omega(n)

Ω(n)\Omega(n)

!!!\infty \star !!!

Why study communication complexity?

Applications

VLSI Time-Space Tradeoff

Proof Complexity

Formula Lower Bounds

Pseudo-randomness...

Space Complexity of Streaming Models

Data Structures

Extension Complexity

Extension Complexity

E.R. Swart. P=NP. Report No. CIS86-02, Department of Computer and Information Science, University of Guelph, Ontario, Canada, 1986

(參考圖片,非當事論文)

Polytope

Intersection of several half-spaces(linear inequalities), and can be expressed as P:={xAxb}P:=\{x|Ax \leq b\}.

face(generated by some supporting hyperplane)

vertex(also a face)

facet(also a face)

Example

By adding auxiliary variables, we obtain an extension of the original polytope.

extension

projection

Example

The extension complexity of a polytope PP is the minimum number of facets(inequalities) achieved by any extension of PP.

extension

projection

What is  xc(TSP(n))xc(TSP(n))?

Nondeterministic Communication Complexity

The setting is the same, A gets xx and B gets yy. However, a person C knows both xx and yy, and C needs to convince A and B that f(x,y)=1f(x,y)=1.

The nondeterministic communication complexity of NEQ is O(logn)O(\log n).

The nondeterministic communication complexity of EQ is Ω(n)\Omega(n).

1-rectangle cover

If there is a 1-rectangle cover of size tt of MfM_f, then there is a nondeterministic protocol of size logt\log t.

If there is a nondeterministic protocol of length cc for ff, then MfM_f can be covered by 2c2^c rectangles.

What is the nondeterministic communication complexity of DISJ?

Extension and Nondeterministic Communication?

xc(P)=rxc(P) = r 

rank+(S)=rrank_+(S) = r, where SS is the slack matrix of PP.

The nondeterministic communication complexity of Face-Vertex(P) is O(logr)O(\log r).

    \implies

Yannakakis' Factorization Theorem

\Leftrightarrow

Slack Matrix

A slack matrix SS of PP is a F×V|F| \times |V| matrix, where FF is the set of faces, and VV is the set of vertices.

Sfv=qfpfTxvS_{fv} = q_f-p_f^Tx_v, where pfTxqfp_f^Tx \leq q_f is the supporting hyperplane of the ff-th face.

Yannakakis' Factorization Theorem-1

xc(P)rank+(S)xc(P) \leq rank_+(S) 

Suppose rank+(S)=rrank_+(S)=r, then we can write S=RTS = RT, where RF×r,Tr×VR_{|F|\times r}, T_{r\times |V|} are nonegative.

We collect the supporting hyperplanes for every face of PP, and obtain AxbA'x \leq b', which is in fact PP itself, and Si =bAviS^i  = b'-A'v_i.

Let Q:={(x,y)Ax+Ry=b,y0}Q:=\{(x,y)|A'x+Ry = b',y \geq 0\}, which has only rr inequalities, and projx(Q)Pproj_x(Q)\subseteq P.

Also Pprojx(Q)P\subseteq proj_x(Q), because for every vertex viv_i of PP, (vi,Ti)Q(v_i,T^i) \in Q.

Yannakakis' Factorization Theorem-2

xc(P)rank+(S)xc(P) \geq rank_+(S) 

Suppose xc(P)=rxc(P)=r, and Q:={(x,y)Cx+Dyd}Q:=\{(x,y)|Cx + Dy \leq d\} is an extension of P:{xAxb}P:\{x|Ax \leq b\}.

Farkas Lemma: Either AxbAx\leq b has a solution xx with xRnx \in \mathbb R^n, or ATy=0A^Ty = 0 has a solution yy with y0,yTb<0y \geq 0,y^Tb < 0.

Each supporting hyperplane generating each face of Q can be expressed as some nonegative linear combination of the facets of QQ.

    \implies

Yannakakis' Factorization Theorem-2

xc(P)rank+(S)xc(P) \geq rank_+(S) 

Suppose xc(P)=rxc(P)=r, and Q:={(x,y)Cx+Dyd}Q:=\{(x,y)|Cx + Dy \leq d\} is an extension of P:{xAxb}P:\{x|Ax \leq b\}.

Each supporting hyperplane generating each face of P can be expressed as some nonegative linear combination of the facets of QQ.

For the ff-th face of PP (generated by pfTxqfp_f^T x \leq q_f), we can find λfR+r\lambda_f \in \mathbb R^r_+, such that λfTC=pT,λfTD=0,λfTd=q\lambda_f^T C = p^T, \lambda_f^T D = 0,\lambda_f^T d = q.

For the vv-th vertex (xvx_v) of PP, we can find (xv,yv)Q(x_v,y_v) \in Q, and let μv=dCxvDyvR+r\mu_v = d-Cx_v-Dy_v \in \mathbb R^r_+.

λfTμv=λfT(dCxvDyv)=qfpfTxv=Sfv\lambda_f^T \mu_v =\lambda_f^T (d - Cx_v - Dy_v)=q_f - p^T_fx_v = S_{fv}

Extension and Nondeterministic Communication

xc(P)=rxc(P) = r 

rank+(S)=rrank_+(S) = r, where SS is the slack matrix of PP.

The nondeterministic communication complexity of FACE-VERTEX(P) is O(logr)O(\log r).

    \implies

Yannakakis' Factorization Theorem

\Leftrightarrow

Example: Correlation Polytope

Define CORR(n)CORR(n) as the convex hull of {y=xTxx{0,1}n}\{y=x^Tx|x \in \{0,1\}^n\}

Can we prove the extension complexity of CORR(n)CORR(n)?

We should consider the nondeterministic communication complexity of FACEVERTEX(CORR(n))FACE-VERTEX(CORR(n)).

The Lemma

For every s{0,1}ns \in \{0,1\}^n, there exist a face fsf_s of CORR(n)CORR(n), such that x{0,1}n\forall x \in \{0,1\}^n

y=xTxfs    sx=1y=x^Tx \in f_s \iff |s \cap x|=1

((si=1xi)1)20((\sum_{s_i = 1} x_i) -1)^2 \geq 0 would be great, but it is not linear.

We change xix_i, xi2x_i^2 into yiiy_{ii}, and change 2xixj2x_ix_j into yij+yjiy_{ij}+y_{ji} to obtain fsf_s.

Reduction?

FACE-VERTEX(CORR(n))

SIZE-OF-INTERSECTION-IS-NOT-ONE(n)

UNIQUE-DISJ(n)

easier

easier

Only consider the faces mentioned in the lemma.

Only consider (x,y)(x,y) such that xy1|x\cap y | \leq 1

Nondeterministic communication complexity of UNIQUE-DISJ!

UNIQUE-DISJ

UNIQUE-DISJ(2)

Still, we want to obtain a lower bound for the size of the 1-rectangle cover of MUNIQUEDISJM_{UNIQUE-DISJ}.

1-rectangle size

Lemma: Each 1-rectangle in MUNIQUEDISJM_{UNIQUE-DISJ} contains at most 2n2^n 1-inputs.

Induction hypothesis: For a 1-rectangle R=A×BR = A\times B, where all xA,yBx\in A,y\in B have 0s in their last nkn-k coordinates, RR contains at most 2k2^k 1-inputs.

k=0k = 0 is obvious, and k=nk=n is what we want.

1-rectangle size

Induction hypothesis: For a 1-rectangle R=A×BR = A\times B, where all xA,yBx\in A,y\in B have 0s in their last nkn-k coordinates, RR contains at most 2k2^k 1-inputs.

For k>0k>0, given RR, we want to construct two sets of 1-inputs S1,S2S_1,S_2, such that

1. rect(S1),rect(S2)rect(S_1),rect(S_2) are 1-rectangles and have 0s in their last nk+1n-k+1 coordinates.

2. the number of 1-entries in S1,S2S_1,S_2 is at least the number of 1-inputs in RR.

We can get rid of of the last nkn-k coordinates.

1-rectangle size

1. rect(S1),rect(S2)rect(S_1),rect(S_2) are 1-rectangles and have 0s in their last nk+1n-k+1 coordinates.

2. The number of 1-entries in S1,S2S_1,S_2 is at least the number of 1-inputs in RR.

Note that  (a1,b1)(a1,b1) is not an 1-input, and (a0,b1),(a1,b0)(a0,b1),(a1,b0) will not be in RR at the same time if (a0,b0)(a0,b0) is an 1-input.

For every 1-input in RR

(a0,b1)    (a0,b1) \implies put it in S1S_1'

(a1,b0)    (a1,b0) \implies put it in S2S_2'

(a0,b0)    (a0,b0) \implies put it in S1S_1' if (a0,b1)R(a_0,b_1)\notin R, put it in S2S_2' if (a1,b0)R(a_1,b_0)\notin R

rect(Si)rect(S_i')s are 1-rectangles. 

xk=0x_k = 0 in rect(S1)rect(S_1'), yk=0y_k = 0 in rect(S2)rect(S_2')

1-rectangle size

Obtain SiS_i by setting yk=0y_k = 0 in rect(S1)rect(S_1') and xk=0x_k = 0 in rect(S2)rect(S_2')

rect(Si)rect(S_i')s are 1-rectangles. 

xk=0x_k = 0 in rect(S1)rect(S_1'), yk=0y_k = 0 in rect(S2)rect(S_2')

Si=Si|S_i| = |S_i'|

1. rect(S1),rect(S2)rect(S_1),rect(S_2) are 1-rectangles and have 0s in their last nk+1n-k+1 coordinates.

2. The number of 1-entries in S1,S2S_1,S_2 is at least the number of 1-inputs in RR.

rect(Si)rect(S_i) have 0s in their last nk+1n-k+1 coordinates.

rect(Si)rect(S_i)s are 1-rectangles. 

Proof Complete!

1-rectangle cover

Lemma: Each 1-rectangle in MUNIQUEDISJ(n)M_{UNIQUE-DISJ(n)} contains at most 2n2^n 1-inputs.

Besides, we know that there are 3n3^n 1-inputs in MUNIQUEDISJ(n)M_{UNIQUE-DISJ(n)}.

Thus, any 1-rectangle cover has size at least (1.5)n(1.5)^n, and the nondeterministic communication complexity of UNIQUE-DISJ(n) would be Ω(n)\Omega(n).

So, finally, 

xc(CORR(n))=2Ω(n)xc(CORR(n)) = 2^{\Omega(n)}

How about TSP(n)TSP(n)?

TSP(n)= conv{χF{0,1}E(Kn)FE(Kn)is a tour of Kn}TSP(n) =  {\rm conv}\{\chi_F \in \{0,1\}^{|E(K_n)|}|F \subseteq E(K_n) \text{is a tour of } K_n \}

There is a face of TSP(cn2)TSP(cn^2) that is an extension of CORR(n)CORR(n).

{xTxx{0,1}n}\{x^Tx|x \in \{0,1\}^n\}

The satisfying assingments of ϕn:=i,j[n],ij(Cij=CiiCjj)\phi_n := \land_{i,j\in[n],i\neq j} (C_{ij} = C_{ii} \land C_{jj})

(O(n2)O(n^2) variables and clauses)

Halmiltonian cycle of some directed graph DnD_n

(O(n2)O(n^2) vertices and edges)

Halmiltonian cycle of some undirected graph GnG_n

(O(n2)O(n^2) vertices and edges)

    \iff

    \iff

    \iff

There is a face of TSP(cn2)TSP(cn^2) that is an extension of CORR(n)CORR(n).

Extension Complexity of TSP(n)

If QQ is an extension of PP, then xc(Q)xc(P)xc(Q)\geq xc(P)

If PP is a face of QQ, then xc(Q)xc(P)xc(Q)\geq xc(P)

(Consider the slack matrix.)

There is a face of TSP(cn2)TSP(cn^2) that is an extension of CORR(n)CORR(n).

The extension complexity of TSP(n)TSP(n) is 2Ω(n)2^{\Omega(\sqrt n)} ><

Q

References

[1] T. Roughgarden. Communication complexity (for algorithm designers). Foundations and Trends in Theoretical Computer Science, 11(3–4):217–404, 2016.

[2] Mathematics and Computation: A Theory Revolutionizing Technology and Science (Princeton University Press, October 2019)

[3] Rao, A., & Yehudayoff, A. (2020). Communication Complexity: And Applications. Cambridge: Cambridge University Press. doi:10.1017/9781108671644

[4] Samuel Fiorini, Serge Massar, Sebastian Pokutta, Hans Raj Tiwary, and Ronald de Wolf. 2015. Exponential Lower Bounds for Polytopes in Combinatorial Optimization. J. ACM 62, 2, Article 17 (May 2015), 23 pages. https://doi.org/10.1145/2716307

Kolmogorov Complexity and the Incompressibility Method

What's the next number?

1,2,4,8,16,1,2,4,8,16,

What's the next number?

1,2,4,8,16,1,2,4,8,16,

71227122?

3232?

What is randomness?

1,2,4,8,16,71221,2,4,8,16,7122

1,2,4,8,16,321,2,4,8,16,32

k = 1
for i in range(5):
  print(k)
  k *= 2
k = 1
for i in range(4):
  print(k)
  k *= 2
print(7122)
A = [1,2,4,8,16,7122]
for i in range(5):
  print(A[i])

Kolmogorov Complexity

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

Kolmogorov Complexity

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

A more formal definition

Cf(x)=min{l(p):f(p)=x}C_f(x) = \min\{l(p):f(p) = x\}

Where ff is some (partial) function.

Kolmogorov Complexity

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

A more formal definition

Cf(x)=min{l(p):f(p)=x}C_f(x) = \min\{l(p):f(p) = x\}

Where ff is some (partial) function.

How should we choose ff? (C++, python, Eric Xiao, ...?)

Kolmogorov Complexity

Maybe ff should be the "minimal" (additively optimal) function:

g,x,Cf(x)Cg(x)+cg\forall g,x, C_f(x) \leq C_g(x) + c_g

Kolmogorov Complexity

Such ff doesn't exist:

Let g(i)=xig(i) = x_i, where Cf(xi)iC_f(x_i) \geq i

Maybe ff should be the "minimal" (additively optimal) function:

g,x,Cf(x)Cg(x)+cg\forall g,x, C_f(x) \leq C_g(x) + c_g

Kolmogorov Complexity

Perhaps we should not consider all functions.

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Let f0f_0 be a function computed by a universal TM (T0T_0) with input  11..110np11..110np (with l(n)l(n) ones ) which simulates TnT_n on input pp

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Let f0f_0 be a function computed by a universal TM (T0T_0) with input  11..110np11..110np (with l(n)l(n) ones ) which simulates TnT_n on input pp

We have Cf0(x)Cfn(x)+2l(n)+1C_{f_0}(x) \leq C_{f_n}(x) + 2l(n)+1

f0f_0 is additively optimal!

Kolmogorov Complexity

C(x)=Cf0(x)=min{l(p):f0(p)=x}C(x) = C_{f_0}(x) = \min\{l(p):f_0(p) = x\}

Kolmogorov Complexity

C(x)=Cf0(x)=min{l(p):f0(p)=x}C(x) = C_{f_0}(x) = \min\{l(p):f_0(p) = x\}

We can view f0f_0 as a special compiler.

Is the compiler of C/Java additively optimal?

Conditional Kolmogorov Complexity

C(xy)=Cf0(xy)=min{l(p):f0(y,p)=x}C(x|y) = C_{f_0}(x|y) = \min\{l(p):f_0(\langle y,p\rangle) = x\}

C(x)=C(xϵ)C(x) = C(x|\epsilon)

We also have f,x,y\forall f,x,y

Cf0(xy)Cf(xy)+cfC_{f_0}(x|y) \leq C_f(x|y) + c_f

Upper bounds

x,C(x)l(x)+c\forall x,C(x) \leq l(x) + c

x,y,C(xy)C(x)+c\forall x,y,C(x|y) \leq C(x)+c

Uncomputability

C(x)C(x) is uncomputable

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

Uncomputability

C(x)C(x) is uncomputable

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

「不能用二十個字以內描述的最小數字」

Uncomputability

C(x)C(x) is uncomputable on any infinite set of points

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

「不能用二十個字以內描述的最小數字」

Incompressibility Theorem \infty \star

Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Incompressibility Theorem \infty \star

Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Because there are only i=0logmc12i=m2c1\sum\limits_{i=0}^{\log m-c-1} 2^i= m2^{-c}-1 programs with length less than logmc\log m-c.

Example - Infinite Primes

Assume that there are only kk primes. Then for each nNn \in \mathbb N, we can write n=i=1kpiαin = \prod\limits_{i=1}^k p_i^{\alpha_i}. Thus we can describe every number in [1,t][1,t] by kloglogtk \log\log t bits......

Incompressity theorem: Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Example - Infinite Primes

Assume that there are only kk primes. Then for each nNn \in \mathbb N, we can write n=i=1kpiαin = \prod\limits_{i=1}^k p_i^{\alpha_i}. Thus we can describe every number in [1,t][1,t] by kloglogtk \log\log t bits......

Incompressity theorem: Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

π(n)Ω(lognloglogn)\pi(n) \in \Omega(\frac{\log n}{\log \log n})

Example - TM complexity

Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Example - TM complexity

Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Assume a TM TT:

  • each state can be described by \ell bits
  • always terminates with the reading head at the rightmost position of the input
  • computes PALINDROME

Example - TM complexity

Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Assume a TM TT:

  • each state can be described by \ell bits
  • always terminates with the reading head at the rightmost position of the input
  • computes PALINDROME

Define the crossing seuence of a grid gg:   the sequence of states of the TM TT when the reading head crosses the right boundary of gg

Example - TM complexity

By the incompressibility theorem, we can find length-nn string xx with C(xT,n)nC(x|T,n) \geq n.

Example - TM complexity

By the incompressibility theorem, we can find length-nn string xx with C(xT,n)nC(x|T,n) \geq n.

Consider running TT on x02nxˉx 0^{2n} \bar x, if the runtime is less than n2\frac{n^2}{\ell}, then there exists a "middle" grid g0g_0 with crossing sequence shorter than n2\frac{n}{2 \ell}.

In fact, given T,nT,n, we can recover xx by the position of g0g_0(n2\frac{n}{2} bits) and the crossing sequence of g0g_0(O(logn)O(\log n) bits).

Example - TM complexity

By the incompressibility theorem, we can find length-nn string xx with C(xT,n)nC(x|T,n) \geq n.

Consider running TT on x02nxˉx 0^{2n} \bar x, if the runtime is less than n2\frac{n^2}{\ell}, then there exists a "middle" grid g0g_0 with crossing sequence shorter than n2\frac{n}{2 \ell}.

In fact, given T,nT,n, we can recover xx by the position of g0g_0(n2\frac{n}{2} bits) and the crossing sequence of g0g_0(O(logn)O(\log n) bits).

C(xT,n)n2+O(logn)C(x|T,n) \leq \frac{n}{2} + O(\log n)

Example - TM complexity

We can recover xx by checking if  the crossing sequence at g0g_0 with input y02nxˉy0^{2n} \bar x is the same as  x02nxˉx0^{2n} \bar x for all y{0,1}ny \in \{0,1\}^n !

Example - TM complexity

We can recover xx by checking if  the crossing sequence at g0g_0 with input y02nxˉy0^{2n} \bar x is the same as  x02nxˉx0^{2n} \bar x for all y{0,1}ny \in \{0,1\}^n !

How about average complexity?

Incompressibility Method

If the statement doesn't hold, then there will be too many compressible objects.

Combinatorial Properties

Some Useful Bounds

nnen1n!nn+1en1\dfrac {n^n} {e^{n - 1} } \le n! \le \dfrac {n^{n + 1} } {e^{n - 1} }

(nk)kCknnkk!<(nek)k\left({\frac{n}{k}}\right)^{k}\leq C^n_k\leq{\frac{n^{k}}{k!}}\lt \left({\frac{n e}{k}}\right)^{k}

Transitive Tournament

Tournament: complete directed graph

 

Transitive tournament: acyclic tournament

 

v(n):v(n): Largest integer such that every tournament on nn nodes contains a subtournament on v(n)v(n) nodes.

Transitive Tournament

Tournament: complete directed graph

 

Transitive tournament: acyclic tournament

 

v(n):v(n): Largest integer such that every tournament on nn nodes contains a subtournament on v(n)v(n) nodes.

logn\lfloor \log n \rfloor \leq v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

Counting

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

v:=2+2lognv:= 2 + \lfloor 2\log n \rfloor

A:A: subset of [n][n] of size vv

σ:\sigma: permutation on [v][v]

ΓA,σ:\Gamma_{A,\sigma}: the set of nn- vertex tournaments of whom AA is a subtournament with "order" σ\sigma

We count the number d(Γ)d(\Gamma') of nn- vertex tournaments with a transitive subtournament of size vv

d(Γ)Aσd(ΓA,σ)=Cvnv!2C2nC2v<2C2nd(\Gamma') \leq \sum_A \sum_\sigma d(\Gamma_{A,\sigma}) = C^n_v v! 2^{C^n_2-C^v_2} < 2^{C^n_2}

(Cvnv!<nv2C2v) (C^n_v v! < n^v \leq 2^{C^v_2})

Probabilistic Method

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

v:=2+2lognv:= 2 + \lfloor 2\log n \rfloor

A:A: subset of [n][n] of size vv

σ:\sigma: permutation on [v][v]

Let TT be the random variable uniformly distributed over all nn-vertex tournaments.

We calculate the probability P(TΓ)P(T\in \Gamma').

P(TΓ)AσP(TΓA,σ)=Cvnv!2C2v<1P(T\in \Gamma') \leq \sum_A \sum_\sigma P(T \in \Gamma_{A,\sigma}) = C^n_v v! 2^{-C^v_2} < 1

(Cvnv!<nv2C2v) (C^n_v v! < n^v \leq 2^{C^v_2})

Incompressibility Method

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

If v(n)v(n) is too large, then we can compress every nn-vertex tournament. (contradicting the incompressibility theorem)

Formally speaking, by the incompressibility method, there is some tournament TT', where C(Tn,p)n(n1)/2C(T'|n,p) \geq n(n-1)/2.

(pp is some decoding program)

However, for every tournament TT, we can describe it by v(n)logn+n(n1)/2v(n)(v(n)1)/2v(n) \lfloor \log n \rfloor+ n(n-1)/2 - v(n)(v(n)-1)/2 bits.

Thus v(n)lognv(n)(v(n)1)/20v(n) \lfloor \log n \rfloor - v(n)(v(n)-1)/2 \geq 0, and

v(n)2logn+1v(n) \leq 2 \lfloor \log n\rfloor + 1.

Incompressibility Method

In fact, by the incompressibility method, we can also prove:

For at least a (11n1 - \frac{1}{n})th fraction of nn-vertex tournaments, the largest transitive subtournament has at most (1+22logn)(1 + 2\lfloor 2\log n\rfloor) nodes for nn large enough.

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

Excerises

1. Give an upper bound for w(n)w(n), the largest integer such that for each nn-vertex tournament TT, there are two disjoint subsets of vertices A,BA,B, each of size w(n)w(n), and A×BTA \times B \in T.

// x,y are n-bit
s = (x ^ y)
c = (x & y)
while c != 0:
  s' = (s ^ c)
  c' = (s & c)
  s = s'
  c = c'

  

2. Prove that the while loop loops at most logn+1\log n + 1 steps in average:

4. What are the differences between the counting argument, the probabilistic method, and the incompressibility method?

5. Ask a question.

3. Prove that almost all nn-vertex labeled trees have maximum degree O(lognloglogn)O(\frac{\log n}{\log \log n}).
W.W. Kirchherr, Kolmogorov complexity and random graphs

Time Complexity

Shell Sort

h1,h2,...,hp=1h_1,h_2,...,h_p=1

In the ii-th pass, divide the list into hih_i sublists (by (pmod  hip \mod h_i)) and do insertion sort.

 

h=[5,3,1]h = [5,3,1]:

Average Time Complexity of Shell Sort

Pratt: Θ(nlog2n)\Theta(n \log^2 n) (worst case also), using hk{2i3j2i3j<N2}h_k \in \{2^i 3 ^j|2^i 3 ^j < \frac{N}{2}\}

Knuth: Θ(n5/3)\Theta(n^{5/3}), using the best choice of hh for p=2p=2

Yao, Janson, Knuth: O(n23/15)O(n^{23/15}), using some choice of hh for p=3p=3

We prove that running shell sort for any pp and (h1,h2,...,hp)(h_1,h_2,...,h_p), the time complexity is Ω(pn1+1/p)\Omega(pn^{1+1/p}).

Proof Outline

Fix some permutation π\pi such that C(πn,A,p)logn!lognC(\pi|n, A,p) \leq \log n! - \log n,

where AA is the algorithm, and pp is some decoding program.

*A fraction of (11n)(1 - \frac{1}{n}) permutations satisfy this condition.

Let mi,km_{i,k} be the distance that the number ii moved in pass kk, and M=i=1nk=1pmi,kM = \sum\limits_{i=1}^n \sum\limits_{k=1}^p m_{i,k}, which is a lower bound of the runtime.

By all mi,km_{i,k}, we can recover π\pi.

If M=o(pn1+1/p)M = o(pn^{1+1/p}), we can express all mi,km_{i,k} with too few bits. (TBD)

Thus the runtime is Ω(pn1+1/p)\Omega(pn^{1+1/p}) for a fraction of (11n)(1-\frac{1}{n}) permutations and so is the average runtime.

Proof Detail

To describe all mi,km_{i,k}, we can descibe

the sum of them (MM, described by O(logn)O(\log n) bits), and

the way of partitioning MM into npnp parts (described by log(HMnp)\log (H^{np}_M) bits).

Thus log(HMnp)+O(logn)logn!logn\log (H^{np}_M) + O(\log n) \geq \log n! - \log n,

log(Cnp1M+np1)logn!+O(logn)\log(C^{M+np-1}_{np-1}) \geq \log n! + O(\log n)

(np1)[log(Mnp1+1)+loge]nlognnloge+O(logn)(np-1)[\log(\frac{M}{np-1}+1)+\log e] \geq n\log n - n \log \mathrm e + O(\log n)

(using Ckn(nek)kC^n_k \leq \left({\frac{n e}{k}}\right)^{k} and n!nnenn! \geq \frac{n^n}{e^n})

Thus M=Ω(pn1+1/p)M = \Omega(pn^{1+1/p}) ???

Regular Language

Regular Language

A language accepted by some DFA.

Pumping Lemma

Pumping Lemma

Might not be easy to use.

Not characterizing.

Example

Is L:={0k1kk1}L := \{0^k 1^k | k \geq 1\} a regular language?

Example

Is L:={0k1kk1}L := \{0^k 1^k | k \geq 1\} a regular language?

Suppose LL is accepted by some DFA AA. Then we can recover kk from AA and qq, the state of AA after processing 0k0^k. Thus for all k1k\geq 1, C(k)O(1)C(k) \leq O(1).

KC-Regularity

Suppose LL is regular, and Lx:={yxyL}L_x := \{y|xy \in L \}. If yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

KC-Regularity

Suppose LL is regular, and Lx:={yxyL}L_x := \{y|xy \in L \}. If yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

*What does "the nn-th string" mean?

More Examples

Is L:={1pp is prime}L := \{1^p |p\text{ is prime}\} a regular language?

Is L:={xxˉwx,w>0}L := \{x \bar x w \large ||x|,|w| > 0\} a regular language?

Regular KC-Characterization

In fact, we have

LL is regular

\Leftrightarrow

Exists cLc_L, such that for all x,nx,n, C(χ1:nx)C(n)+cLC(\chi^x_{1:n}) \leq C(n) + c_L

For all x,nx,n, if yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

\Leftrightarrow

Regular KC-Characterization

In fact, we have

LL is regular

\Leftrightarrow

Exists cLc_L, such that for all x,nx,n, C(χ1:nx)C(n)+cLC(\chi^x_{1:n}) \leq C(n) + c_L

For all x,nx,n, if yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

\Leftrightarrow

Thus, (in theory), we can prove non-regularity of every non-regular language and regularity of every regular language.

Last Example

Is L:={xxi=2k+1}L := \{x |\sum x_i = 2k+1\} a regular language?

Last Example

Is L:={xxi=2k+1}L := \{x |\sum x_i = 2k+1\} a regular language?

There exists cLc_L, such that for all x,nx,n, C(χ1:nx)C(n)+cLC(\chi^x_{1:n}) \leq C(n) + c_L.

Excerises

1. Prove (or disprove) that L:={0i1jgcd(i,j)=1}L:=\{0^i 1^j|\gcd(i,j) = 1\} is regular.

2. Alice and Bob are playing the number guessing game. Alice chooses some x[n]x \in [n], and Bob can ask her yes/no questions. If Alice can lie once, what is the minimum number cc where Bob can always guess the right answer after asking cc questions?

5. Why is computability important?

6. Ask a question.

4. Prove any (nontrivial?) statement via the incompressibility method.

3. An nni (nearest-neighbor interchange) operation can be done on neighboring internal nodes on an evolutionary tree. Let f(A,B)f(A,B) denote the nni-distance between two nn-node evolutionary trees. Prove that maxA,Bf(A,B)=Ω(nlogn)\max_{A,B} f(A,B) = \Omega(n \log n). (In fact, Θ(nlogn)\Theta(n \log n).)

References

[1] Wikipedia contributors. (2023, March 5). Incompressibility method. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Incompressibility_method&oldid=1143081380

[2] Wikipedia contributors. (2023, August 22). Kolmogorov complexity. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Kolmogorov_complexity&oldid=1171617849

[3] Ming Li , Paul Vitányi. An Introduction to Kolmogorov Complexity and Its Applications, 2008. 

A Glimpse into

Algorithmic Game Theory

2024.2.15

Other Considerd Topics - Spectral Graph Theory

  • Spectrum of Adjacency Matrix/Laplacian
  • Expander Graphs/PSRGs
  • Expander Codes
  • Nearly-linear time Laplacian solvers

Other Considerd Topics - Sum of Squares Hierachy

  • Semidefinite Programming/Polynomial Optimization?
  • polynomial-time 0.878-approximation algorithm for max cut (and why this might be optimal)

Today's Topic - Algorithmic Game Theory

  • Auction Analysis
  • Price of Anarchy (PoA, 最壞均衡與最佳解比)

What is a Game?

A game is defined by 

1. A set of nn players

2. The pure strategy set for each player: S1,S2,...,SnS_1,S_2,...,S_n

3. The payoff function for each player: u1,u2,...,un:S1×S2×...×SnRu_1, u_2, ..., u_n : S_1 \times S_2 \times ... \times S_n \to \mathbb R

https://www.youtube.com/watch?app=desktop&v=-1GDMXoMdaY

Equilibriums

A pure strategy profile sS1×S2×...×Sns^* \in S_1\times S_2\times ... \times S_n is a Pure Nash Equilibrium (PNE) if

ui(s)ui(si,si)i[n] and siSi{ u}_{i}\left(s^*\right)\,\ge\,{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)\, \forall i\in [n] \text{ and }s_i \in S_i

A mixed strategy profile pΔ(S1)×Δ(S2)×...×Δ(Sn)p^* \in \Delta(S_1)\times \Delta(S_2)\times ... \times \Delta(S_n) is a Mixed Nash Equilibrium (MNE) if

Esp[ui(s)]Esipi[ui(si,si)]i[n] and siSi\mathbb E_{s^* \sim p^*}[{ u}_{i}\left(s^*\right)]\,\ge\,\mathbb E_{s_{-i}^*\sim p^*_{-i}}[{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)]\, \forall i\in [n] \text{ and }s_i \in S_i

Equilibriums

A pure strategy profile sS1×S2×...×Sns^* \in S_1\times S_2\times ... \times S_n is a Pure Nash Equilibrium (PNE) if

ui(s)ui(si,si)i[n] and siSi{ u}_{i}\left(s^*\right)\,\ge\,{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)\, \forall i\in [n] \text{ and }s_i \in S_i

A mixed strategy profile pΔ(S1)×Δ(S2)×...×Δ(Sn)p^* \in \Delta(S_1)\times \Delta(S_2)\times ... \times \Delta(S_n) is a Mixed Nash Equilibrium (MNE) if

Esp[ui(s)]Esipi[ui(si,si)]i[n] and siSi\mathbb E_{s^* \sim p^*}[{ u}_{i}\left(s^*\right)]\,\ge\,\mathbb E_{s_{-i}^*\sim p^*_{-i}}[{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)]\, \forall i\in [n] \text{ and }s_i \in S_i

Nash's theorem:  There is at least one MNE for each game.

Q: What is an MNE for two-player rock paper scissors?

Equilibriums

A pure strategy profile sS1×S2×...×Sns^* \in S_1\times S_2\times ... \times S_n is a Pure Nash Equilibrium (PNE) if

ui(s)ui(si,si)i[n] and siSi{ u}_{i}\left(s^*\right)\,\ge\,{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)\, \forall i\in [n] \text{ and }s_i \in S_i

A mixed strategy profile pΔ(S1)×Δ(S2)×...×Δ(Sn)p^* \in \Delta(S_1)\times \Delta(S_2)\times ... \times \Delta(S_n) is a Mixed Nash Equilibrium (MNE) if

Esp[ui(s)]Esipi[ui(si,si)]i[n] and siSi\mathbb E_{s^* \sim p^*}[{ u}_{i}\left(s^*\right)]\,\ge\,\mathbb E_{s_{-i}^*\sim p^*_{-i}}[{u}_{i}\left(s_{i}\,,\,s_{-i}^{\ast}\right)]\, \forall i\in [n] \text{ and }s_i \in S_i

Nash's theorem:  There is at least one MNE for each game.

Q: Why are equilibriums important?

What is Economics?

What is Economics?

  • 經世濟民
  • 尋找最有效率分配資源的方式

An Efficient Auction Mechanism:

Single-Item Second-Price Auction

1. Each of the nn bidders has a private valuation viv_i of the item.

2. Each bidder submits its bid bib_i to the auctioneer simultaneously.

3. The auctioneer gives the item to the kk-th bidder where bk=maxibib_k=\max_{i} b_i and charges p(b)=maxjkbjp(b) = \max_{j\neq k} b_j

An Efficient Auction Mechanism:

Single-Item Second-Price Auction

1. Each of the nn bidders has a private valuation viv_i of the item.

2. Each bidder submits its bid bib_i to the auctioneer simultaneously.

3. The auctioneer gives the item to the kk-th bidder where bk=maxibib_k=\max_{i} b_i and charges p(b)=maxjkbjp(b) = \max_{j\neq k} b_j

  • Bidding truthfully (bi=vi)(b_i = v_i) maximizes the utility ui(b)=(vip(b))[i=k]u_i(b) = (v_i -p(b)) \cdot [ i = k] for every bidder.
  • The social welfare SW(b)=vk\mathrm{SW}(b) = v_{k} is always maximized when all bidders bid truthfully.

How about Single-Item First-Price Auctions?

1. Each of the nn bidders has a private valuation viv_i of the item.

2. Each bidder submits its bid bib_i to the auctioneer simultaneously.

3. The auctioneer gives the item to the kk-th bidder where bk=maxibib_k=\max_{i} b_i and charges bkb_k

How about Single-Item First-Price Auctions?

1. Each of the nn bidders has a private valuation viv_i of the item.

2. Each bidder submits its bid bib_i to the auctioneer simultaneously.

3. The auctioneer gives the item to the kk-th bidder where bk=maxibib_k=\max_{i} b_i and charges bkb_k

Q:What  will the bidders' strategy be? Is the welfare always maximzied?

How about Single-Item First-Price Auctions?

1. Each of the nn bidders has a private valuation viv_i of the item.

2. Each bidder submits its bid bib_i to the auctioneer simultaneously.

3. The auctioneer gives the item to the kk-th bidder where bk=maxibib_k=\max_{i} b_i and charges bkb_k

Questions we will consider:

  • How can we define the efficiency(inefficiency) of the first-price auction?
  • How can we find/prove the efficiency of the first-price auction?
  • Other kinds of auctions? More than one item?

Model:

Games(Auctions) of Incomplete Information

Assumption:

The valuation viv_i of each bidder is drawn independently from some publicly-known distribution Fi\cal F_i. (Fi\cal F_i Corresponds to the common beliefs that bidders have about everyone’s valuations.)

A strategy of the ii-th bidder is a function si:SUPP(Fi)R+s_i: \mathrm{SUPP}(\cal F_i) \to \mathbb R^+. (When my valuation is viv_i , I will bid si(vi)s_i(v_i).)

A strategy profile ss is a Bayes-Nash Equilibrium(BNE) if i[n],viSUPP(Fi)\forall i \in [n], v_i \in \mathrm{SUPP}(\cal F_i)

EviFi[ui(s(vi,vi))]EviFi[ui(si(vi),r)]\mathbb E_{v_{-i} \sim \cal F_{-i}}[u_i(s(v_{-i},v_i))] \geq \mathbb E_{v_{-i} \sim \cal F_{-i}}[u_i(s_{-i}(v_{-i}),r')], rR+\forall r' \in \mathbb R^+

Model:

Games(Auctions) of Incomplete Information

Assumption:

The valuation viv_i of each bidder is drawn independently from some publicly-known distribution Fi\cal F_i. (Fi\cal F_i Corresponds to the common beliefs that bidders have about everyone’s valuations.)

A strategy of the ii-th bidder is a function si:SUPP(Fi)R+s_i: \mathrm{SUPP}(\cal F_i) \to \mathbb R^+. (When my valuation is viv_i , I will bid si(vi)s_i(v_i).)

A strategy profile ss is a Bayes-Nash Equilibrium(BNE) if i[n],viSUPP(Fi)\forall i \in [n], v_i \in \mathrm{SUPP}(\cal F_i)

EviFi[ui(s(vi,vi))]EviFi[ui(si(vi),r)]\mathbb E_{v_{-i} \sim \cal F_{-i}}[u_i(s(v_{-i},v_i))] \geq \mathbb E_{v_{-i} \sim \cal F_{-i}}[u_i(s_{-i}(v_{-i}),r')], rR+\forall r' \in \mathbb R^+

Q : The mixed version?

Example:
Two bidders, symmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1,v2v_1, v_2 i.i.d. follow U[0,1]U[0,1]. Please find a (pure) BNE.

Example:
Two bidders, symmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1,v2v_1, v_2 i.i.d. follow U[0,1]U[0,1]. Please find a (pure) BNE.

Hint: The two bidders' strategies are identical and monotone.

Example:
Two bidders, symmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1,v2v_1, v_2 i.i.d. follow U[0,1]U[0,1]. Please find a (pure) BNE.

Q:

Is the BNE easy to solve?

Does the BNE maximize the welfare?

Is the auction efficient in this case?

Example:
Two bidders, asymmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1U[0,1],v2U[0,2]v_1 \sim U[0,1], v_2\sim U[0,2] where v1v2v_1 \perp v_2. Please find the BNE.

Example:
Two bidders, asymmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1U[0,1],v2U[0,2]v_1 \sim U[0,1], v_2\sim U[0,2] where v1v2v_1 \perp v_2.

s1(v1)=  43v1(113v124)s_{1}(v_{1})=\;\frac{4}{3v_{1}}\left(1-\sqrt{1-\frac{3v_{1}^{2}}{4}}\right)

s2(v2)=  43v2(1+3v2241)s_{2}(v_{2})=\;\frac{4}{3v_{2}}\left(\sqrt{1+\frac{3v_{2}^{2}}{4}}-1\right)

Example:
Two bidders, asymmetric uniform valuation

Q:

There are two bidders in a first-price auction, and v1U[0,1],v2U[0,2]v_1 \sim U[0,1], v_2\sim U[0,2] where v1v2v_1 \perp v_2.

Q:

Is the BNE easy to solve?

Does the BNE maximize the welfare?

Is the auction efficient in this case?

s1(v1)=  43v1(113v124)s_{1}(v_{1})=\;\frac{4}{3v_{1}}\left(1-\sqrt{1-\frac{3v_{1}^{2}}{4}}\right)

s2(v2)=  43v2(1+3v2241)s_{2}(v_{2})=\;\frac{4}{3v_{2}}\left(\sqrt{1+\frac{3v_{2}^{2}}{4}}-1\right)

Some notations

Valuation profile:  v=(v1,v2,...,vn)v = (v_1, v_2, ..., v_n)

Bid profile:  b=(b1,b2,...,bn)b = (b_1, b_2, ..., b_n)

Whether or not bidder ii is the winner: xi(b)x_i(b)

Selling price: p(b)=maxibip(b) = \max_i b_i

Utility: ui(b,vi)=(vibi)xi(b)u_i(b,v_i)=(v_i-b_i) \cdot x_i(b)

Strategy profile: s=(s1,s2,...,sn):SUPP(F)Rns = (s_1, s_2, ..., s_n): \mathrm{SUPP}(\cal F) \to \mathbb R^n

Social Welfare: SW(b;v)=i=1nvixi(b)\mathrm{SW}(b;v) = \sum\limits_{i=1}^n v_i \cdot x_i(b)

Whether or not bidder ii has the highest valuation: xi(v)x_i^*(v)

Optimal social welfare: OPT(v)=maxivi=i=1nvixi(v)\mathrm{OPT}(v) = \max_i v_i = \sum\limits_{i=1}^n v_i \cdot x_i^*(v)

\infty \star Price of Anarchy

Social Welfare: SW(b;v)=i=1nvixi(b)\mathrm{SW}(b;v) = \sum\limits_{i=1}^n v_i \cdot x_i(b)

Optimal social welfare: OPT(v)=maxivi\mathrm{OPT}(v) = \max_i v_i =i=1nvixi(v) = \sum\limits_{i=1}^n v_i \cdot x_i^*(v)

Price of Anarchy:  PoA=mins is a BNEEvF[SW(s(v);v)]EvF[OPT(v)]\mathrm{PoA} = \min_{s \text{ is a BNE}} \frac{\mathbb E_{v \sim \cal F}[\mathrm{SW}(s(v);v)]}{\mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)]}

\infty \star Price of Anarchy

Social Welfare: SW(b;v)=i=1nvixi(b)\mathrm{SW}(b;v) = \sum\limits_{i=1}^n v_i \cdot x_i(b)

Optimal social welfare: OPT(v)=maxivi\mathrm{OPT}(v) = \max_i v_i =i=1nvixi(v) = \sum\limits_{i=1}^n v_i \cdot x_i^*(v)

Price of Anarchy:  PoA=mins is a BNEEvF[SW(s(v);v)]EvF[OPT(v)]\mathrm{PoA} = \min_{s \text{ is a BNE}} \frac{\mathbb E_{v \sim \cal F}[\mathrm{SW}(s(v);v)]}{\mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)]}

High PoA     \implies efficient auction?

Low PoA     \implies inefficient auction?

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

ui(bi,vi2;vi)(12vip(b))u_i(b_{-i},\frac{v_i}{2};v_i)\geq (\frac{1}{2} v_i - p(b))

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

ui(bi,vi2;vi)(12vip(b))xi(v)u_i(b_{-i},\frac{v_i}{2};v_i)\geq (\frac{1}{2} v_i - p(b))\cdot x_i^*(v)

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

ui(bi,vi2;vi)(12vip(b))xi(v)u_i(b_{-i},\frac{v_i}{2};v_i)\geq (\frac{1}{2} v_i - p(b))\cdot x_i^*(v)

2. For every BNE ss, bidder ii, and its valuation viv_i

Evi[ui(s(v);vi)]Evi[ui(vi2,si(vi);vi)].\mathbb{E}_{v_{-i}}\left[u_{i}({s}({v});v_{i})\right]\geq\mathbb{E}_{{v}_{-i}}\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii, and its valuation viv_i

Evi[ui(s(v);vi)]Evi[ui(vi2,si(vi);vi)].\mathbb{E}_{v_{-i}}\left[u_{i}({s}({v});v_{i})\right]\geq\mathbb{E}_{{v}_{-i}}\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii,

EviEvi[ui(s(v);vi)]EviEvi[ui(vi2,si(vi);vi)].\mathbb E_{v_i}\mathbb{E}_{v_{-i}}\left[u_{i}({s}({v});v_{i})\right]\geq\mathbb E_{v_i}\mathbb{E}_{{v}_{-i}}\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii,

Ev[ui(s(v);vi)]Ev[ui(vi2,si(vi);vi)].\mathbb{E}_{v}\left[u_{i}({s}({v});v_i)\right]\geq\mathbb{E}_{{v}}\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii,

Evi[ui(s(v);vi)]Evi[ui(vi2,si(vi);vi)].\mathbb{E}_{v} \sum_i\left[u_{i}({s}({v});v_i)\right]\geq\mathbb{E}_{{v}} \sum_i\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii,

EvSW(s(v);v)p(s(v))=Evi[ui(s(v);vi)]Evi[ui(vi2,si(vi);vi)].\mathbb E_v \mathrm{SW}(s(v);v) - p(s(v))=\mathbb{E}_{v} \sum_i\left[u_{i}({s}({v});v_i)\right]\geq\mathbb{E}_{{v}} \sum_i\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

Proof: We will proof the 12\frac{1}{2} PoA bound. There are two main inequalities.

1. For every valuation profile vv, bid profile bb, and bidder ii:

iui(bi,vi2;vi)i(12vip(b))xi(v)=12OPT(v)p(b) \sum_i u_i(b_{-i},\frac{v_i}{2};v_i)\geq \sum_i(\frac{1}{2} v_i - p(b))\cdot x_i^*(v) = \frac{1}{2}\mathrm{OPT}(v)-p(b)

2. For every BNE ss, bidder ii,

EvSW(s(v);v)p(s(v))=Evi[ui(s(v);vi)]Evi[ui(vi2,si(vi);vi)].\mathbb E_v \mathrm{SW}(s(v);v) - p(s(v))=\mathbb{E}_{v} \sum_i\left[u_{i}({s}({v});v_i)\right]\geq\mathbb{E}_{{v}} \sum_i\left[u_{i}(\frac{v_i}{2},{s}_{-i}({v}_{-i});v_{i})\right].

    PoA12\implies \mathrm{PoA} \geq \frac{1}{2}

Theorem 1

The PoA of first-price single-item auction is at least 11e0.631-\frac{1}{e} \approx 0.63

* The 11e1-\frac{1}{e} bound can be proved similarly.

* The bound holds even if the valuations are correlated!

More Items? Simultaneous First-price Auctions

1. Each of the nn bidders has a private valuation vij(i[n],j[m])v_{ij}(i \in [n], j \in [m]) to each of the mm item. (Assume viFiv_i \sim \cal F_i)

2. Each bidder submits its bid bi=(bi1,bi2,...,bim)b_i = (b_{i1},b_{i2}, ..., b_{im}) to the auctioneer simultaneously.

3. The auctioneer gives the jj-th item to the kk-th bidder where bkj=maxibijb_{kj}=\max_{i} b_{ij} and charges bkjb_{kj}

More Items? Simultaneous First-price Auctions

1. Each of the nn bidders has a private valuation vij(i[n],j[m])v_{ij}(i \in [n], j \in [m]) to each of the mm item. (Assume viFiv_i \sim \cal F_i)

2. Each bidder submits its bid bi=(bi1,bi2,...,bim)b_i = (b_{i1},b_{i2}, ..., b_{im}) to the auctioneer simultaneously.

3. The auctioneer gives the jj-th item to the kk-th bidder where bkj=maxibijb_{kj}=\max_{i} b_{ij} and charges bkjb_{kj}

Denote by Si(b)S_i(b) the set of items given to bidder ii.

We assume the bidders are unit-demand:

ui(b;vi)=maxjSi(b)vijjSi(b)biju_i(b;v_i) = \max_{j \in S_i(b)} v_{ij} - \sum_{j \in S_i(b)} b_{ij}

More Items? Simultaneous First-price Auctions

1. Each of the nn bidders has a private valuation vij(i[n],j[m])v_{ij}(i \in [n], j \in [m]) to each of the mm item. (Assume viFiv_i \sim \cal F_i)

2. Each bidder submits its bid bi=(bi1,bi2,...,bim)b_i = (b_{i1},b_{i2}, ..., b_{im}) to the auctioneer simultaneously.

3. The auctioneer gives the jj-th item to the kk-th bidder where bkj=maxibijb_{kj}=\max_{i} b_{ij} and charges bkjb_{kj}

Denote by Si(b)S_i(b) the set of items given to bidder ii.

We assume the bidders are unit-demand:

ui(b;vi)=maxjSi(b)vijjSi(b)biju_i(b;v_i) = \max_{j \in S_i(b)} v_{ij} - \sum_{j \in S_i(b)} b_{ij}

Maximum weight matching?

Is there a PoA bound for

S1A with unit-demand bidders?

Theorem 2

The PoA of S1A with unit demand bidders is at least 11e0.631-\frac{1}{e} \approx 0.63

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

We will first prove the complete-information version:

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Theorem 2

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

Denote by j(i)j^*(i), the item bidder ii gets in some fixed optimal allocation. Define bidder ii's deviation bib_i^* as bidding vij(i)2\frac{v_{ij^*(i)}}{2} on j(i)j^*(i) and 00 on other items.

Theorem 2

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

Denote by j(i)j^*(i), the item bidder ii gets in some fixed optimal allocation. Define bidder ii's deviation bib_i^* as bidding vij(i)2\frac{v_{ij^*(i)}}{2} on j(i)j^*(i) and 00 on other items.

Like the proof of theorem 1, we have for every bid profile bb,

ui(bi,bi;υi)vij(i)2pj(i)(b)u_{i}(b_{i}^{*},{b}_{-i}^{};\upsilon_{i})\,\ge\,\frac{v_{i j^{*}(i)}}{2}\,-\,p_{j^*(i)}(b)

Theorem 2

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

Like the proof of theorem 1, we have for every bid profile bb,

ui(bi,bi;υi)vij(i)2pj(i)(b)u_{i}(b_{i}^{*},{b}_{-i}^{};\upsilon_{i})\,\ge\,\frac{v_{i j^{*}(i)}}{2}\,-\,p_{j^*(i)}(b)

Thus i[n]ui(bi,bi;vi)12OPT(v)j[m]pj(b).\sum_{i\in[n]}u_{i}(b_{i}^{*},{ b}_{-i};v_{i})\geq\frac{1}{2}\mathrm{OPT}({ v})-\sum_{j\in[m]}p_{j}({ b}).

Theorem 2

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

Like the proof of theorem 1, we have for every bid profile bb,

ui(bi,bi;υi)vij(i)2pj(i)(b)u_{i}(b_{i}^{*},{b}_{-i}^{};\upsilon_{i})\,\ge\,\frac{v_{i j^{*}(i)}}{2}\,-\,p_{j^*(i)}(b)

Thus i[n]ui(bi,bi;vi)12OPT(v)j[m]pj(b).\sum_{i\in[n]}u_{i}(b_{i}^{*},{ b}_{-i};v_{i})\geq\frac{1}{2}\mathrm{OPT}({ v})-\sum_{j\in[m]}p_{j}({ b}).

Like the proof of theorem 1, using the definition of MNE, we can have PoA 12\geq \frac{1}{2}.

Theorem 2

Every MNE of the S1A with unit-demand bidders and fixed valuations (Fi\cal F_i is a degenarate distribution at viv_i) achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

Like the proof of theorem 1, we have for every bid profile bb,

ui(bi,bi;υi)vij(i)2pj(i)(b)u_{i}(b_{i}^{*},{b}_{-i}^{};\upsilon_{i})\,\ge\,\frac{v_{i j^{*}(i)}}{2}\,-\,p_{j^*(i)}(b)

Thus i[n]ui(bi,bi;vi)12OPT(v)j[m]pj(b).\sum_{i\in[n]}u_{i}(b_{i}^{*},{ b}_{-i};v_{i})\geq\frac{1}{2}\mathrm{OPT}({ v})-\sum_{j\in[m]}p_{j}({ b}).

Like the proof of theorem 1, using the definition of MNE, we can have PoA 12\geq \frac{1}{2}.

Can we use the same proof for the incomplete-information case?

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We fix some BNE, where bidder ii samples its bid from the distribution Di(vi)D_i(v_i) when its valuation is viv_i. Denote by Gi\cal G_i the distribution of bidder ii's bid, and G=G1×G2×...×Gn\cal G = \cal G_1 \times \cal G_2 \times ... \times \cal G_n.

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We fix some BNE, where bidder ii samples its bid from the distribution Di(vi)D_i(v_i) when its valuation is viv_i. Denote by Gi\cal G_i the distribution of bidder ii's bid, and G=G1×G2×...×Gn\cal G = \cal G_1 \times \cal G_2 \times ... \times \cal G_n.


We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE, for every ii and viv_i:

EbiD(vi),biGi[ui(b;vi)]\mathbb E_{b_i \sim D(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b;v_i)] \geq EbiD(vi),biGi[ui(bi,bi;vi)]\mathbb E_{b_i' \sim D'(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b_{-i},b_i';v_i)]

=EviFi,biGi[ui(bi,bi(v);vi)]=\mathbb E_{v_{-i} \sim \cal F_{-i}, b_{-i} \sim \cal G_{-i}}[u_i(b_{-i},b_i^*(v);v_i)]

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE, for every ii and viv_i:

EbiD(vi),biGi[ui(b;vi)]\mathbb E_{b_i \sim D(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b;v_i)] \geq EviFi,biGi[ui(bi,bi(v);vi)]\mathbb E_{v_{-i} \sim \cal F_{-i}, b_{-i} \sim \cal G_{-i}}[u_i(b_{-i},b_i^*(v);v_i)]

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE, for every ii:

EviFi,biD(vi),biGi[ui(b;vi)]\mathbb E_{v_i \sim \cal F_i,b_i \sim D(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b;v_i)] \geq EvF,bG[ui(bi,bi(v);vi)]\mathbb E_{v \sim \cal F, b \sim \cal G}[u_i(b_{-i},b_i^*(v);v_i)]

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE:

iEviFi,biD(vi),biGi[ui(b;vi)]\sum_i \mathbb E_{v_i \sim \cal F_i,b_i \sim D(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b;v_i)] \geq EvF,bG[iui(bi,bi(v);vi)]\mathbb E_{v \sim \cal F, b \sim \cal G}[\sum_i u_i(b_{-i},b_i^*(v);v_i)]

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE:

iEviFi,biD(vi),biGi[ui(b;vi)]\sum_i \mathbb E_{v_i \sim \cal F_i,b_i \sim D(v_i), b_{-i} \sim \cal G_{-i}}[u_i(b;v_i)] \geq EvF,bG[iui(bi,bi(v);vi)]\mathbb E_{v \sim \cal F, b \sim \cal G}[\sum_i u_i(b_{-i},b_i^*(v);v_i)]

Last Proof:

iui(bi,bi;vi)12OPT(v)j[m]pj(b).\sum_{i}u_{i}(b_{i}^{*},{ b}_{-i};v_{i})\geq\frac{1}{2}\mathrm{OPT}({ v})-\sum_{j\in[m]}p_{j}({ b}).

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE:

EvF,biD(vi)i[iui(b;vi)]\mathbb E_{v \sim \cal F,b_i \sim D(v_i)\forall i}[\sum_i u_i(b;v_i)] \geq 12EvF[OPT(v)]EbG[R(b)]\frac{1}{2} \mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)] - \mathbb E_{b \sim \cal G}[R(b)]

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE:

EvF,biD(vi)i[(iui(b;vi))+R(b)]\mathbb E_{v \sim \cal F,b_i \sim D(v_i)\forall i}[(\sum_i u_i(b;v_i))+R(b)] \geq 12EvF[OPT(v)]\frac{1}{2} \mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)]

Expected social welfare under the BNE.

Theorem 2

Every BNE of the S1A with unit-demand bidders and independent valuations achieves expected social welfare at least 11e1 − \frac{1}{e} times the expected optimal welfare.

Proof: (We will still prove the 12\frac{1}{2} PoA bound only.)

We will use the randomized deviation biDi(vi)b_i' \sim D_i'(v_i): sample viv_{-i} from Fi\cal F_{-i} and then performs bi(v)b_i^*(v).

By the definition of BNE:

EvF,biD(vi)i[(iui(b;vi))+R(b)]\mathbb E_{v \sim \cal F,b_i \sim D(v_i)\forall i}[(\sum_i u_i(b;v_i))+R(b)] \geq 12EvF[OPT(v)]\frac{1}{2} \mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)]

Expected social welfare under the BNE.

DONE!

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}

The proof techniques are similar:

1. Choose some deviation bib_i'.

2. Find some lower bound for iui(bi,bi;vi)\sum_i u_i(b_{-i},b_i';v_i) where the lower bound is related with OPT(v)\mathrm{OPT}(v).

3. By the definition of BNE, the lower bound in 2. is also a lower bound for the expected social welfare under the BNE.

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}

The proof techniques are similar:

Can the proof technique be used to prove PoA for other auctions?

Can we turn the proof technique into some general theorem?

General Auction Mechanisms

1. The auctioneer receives an action aiAia_i\in \cal A_i from each player

2. Given a=(a1,a2,...,an)a = (a_1, a_2, ..., a_n), the auctioneer decides an outcome o(a)Oo(a) \in \cal O

3. Each player derives some utility ui(a;vi):=ui(o(a);vi)u_i(a;v_i) := u_i(o(a);v_i), where viv_i is the valuation of the player

4. R(a):=R(o(a))R(a) := R(o(a)) is the revenue of the auctioneer

*The BNE, social welfare, optimal welfare, PoA can also be defined.

General Auction Mechanisms

1. The auctioneer receives an action aiAia_i\in \cal A_i from each player

2. Given a=(a1,a2,...,an)a = (a_1, a_2, ..., a_n), the auctioneer decides an outcome o(a)Oo(a) \in \cal O

3. Each player derives some utility ui(a;vi):=ui(o(a);vi)u_i(a;v_i) := u_i(o(a);v_i), where viv_i is the valuation of the player

4. R(a):=R(o(a))R(a) := R(o(a)) is the revenue of the auctioneer

Examples:
- combinatorial auctions: mm items and nn players, each player has a valuation for each of the 2m2^m subsets of items

    - single item auctions (first price, second price, all pay auctions, ...)

    - S1A with unit-demand bidders

General Auction Mechanisms

1. The auctioneer receives an action aiAia_i\in \cal A_i from each player

2. Given a=(a1,a2,...,an)a = (a_1, a_2, ..., a_n), the auctioneer decides an outcome o(a)Oo(a) \in \cal O

3. Each player derives some utility ui(a;vi):=ui(o(a);vi)u_i(a;v_i) := u_i(o(a);v_i), where viv_i is the valuation of the player

4. R(a):=R(o(a))R(a) := R(o(a)) is the revenue of the auctioneer

Examples:
- combinatorial auctions

- public good: There is some cost cc publicly known, and each of the nn players has its own valuation viv_i.

Smooth Auction

For λ0,μ1\lambda \geq 0, \mu \geq 1, an auction is (λ,μ)(\lambda, \mu)-smooth if for every valuation profile vv, there exists action distributions D1(v),D2(v),...,Dn(v)D_1^*(v), D_2^*(v), ..., D_n^*(v) such that for every action profile aa:

iEaiDi(v)[ui(ai,ai;vi)]λOPT(v)μR(a).\sum_{i}\mathbb{E}_{a_{i}^{*}\sim D_{i}^{*}({v})}\left[u_{i}(a_{i}^{*},{a}_{-i};v_{i})\right]\geq\lambda\mathrm{OPT}({v})-\mu R({a}).

Smooth Auction

For λ0,μ1\lambda \geq 0, \mu \geq 1, an auction is (λ,μ)(\lambda, \mu)-smooth if for every valuation profile vv, there exists action distributions D1(v),D2(v),...,Dn(v)D_1^*(v), D_2^*(v), ..., D_n^*(v) such that for every action profile aa:

iEaiDi(v)[ui(ai,ai;vi)]λOPT(v)μR(a).\sum_{i}\mathbb{E}_{a_{i}^{*}\sim D_{i}^{*}({v})}\left[u_{i}(a_{i}^{*},{a}_{-i};v_{i})\right]\geq\lambda\mathrm{OPT}({v})-\mu R({a}).

Almost same proof of Theorem 2

(complete information version)

If an auction is (λ,μ)(\lambda, \mu)-smooth, then for every valuation profile vv, every complete-information MNE of the auction has expected welfare at least λμOPT(v)\frac{\lambda}{\mu} \cdot \mathrm{OPT}(v).

Smooth Auction

For λ0,μ1\lambda \geq 0, \mu \geq 1, an auction is (λ,μ)(\lambda, \mu)-smooth if for every valuation profile vv, there exists action distributions D1(v),D2(v),...,Dn(v)D_1^*(v), D_2^*(v), ..., D_n^*(v) such that for every action profile aa:

iEaiDi(v)[ui(ai,ai;vi)]λOPT(v)μR(a).\sum_{i}\mathbb{E}_{a_{i}^{*}\sim D_{i}^{*}({v})}\left[u_{i}(a_{i}^{*},{a}_{-i};v_{i})\right]\geq\lambda\mathrm{OPT}({v})-\mu R({a}).

Almost same proof of Theorem 2

(incomplete information version)

\infty \star If an auction is (λ,μ)(\lambda, \mu)-smooth, then for all valuation distributions F1,F2,...,Fn\cal F_1, \cal F_2, ..., \cal F_n, every BNE of the auction has expected welfare at least λμEvF[OPT(v)]\frac{\lambda}{\mu} \cdot \mathbb E_{v \sim \cal F}[\mathrm{OPT}(v)].

Excercises!

1. Prove the 12\frac{1}{2} PoA bound for single-item all-pay auction. 

2. Prove the (11e)(1-\frac{1}{e}) PoA bound for single-item first-price auction.

3*. Prove a PoA bound for the single-item auction where the two bidders with the highest bids needs to pay and the bidder with the highest bid gets the item (or prove that no PoA bound holds).

4. 剛剛講的那些東西"algorithmic"在哪裡?

5. When were the mentioned results derived?
     1920 (A) 1940 (B) 1960 (C) 1980 (D) 2000 (E) 2020

 

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.

Critique: The strategies of the players do not always form BNE in reality!

No-Regret Learning

In this scenario, we assume that the same nn bidders participate in the same auction for TT rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid. 

No-Regret Learning

In this scenario, we assume that the same nn bidders participate in the same auction for TT rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid. 

We assume that all players choose their bidding stategies aita^t_i to have o(T)o(T) regret (no regret), where the regret is defined as 

RT=(maxat=1Tuit(a))(t=1Tuit(ait) )R_T = \left (\max_{a'} \sum\limits_{t = 1}^T u^t_i(a')\right) - \left( \sum\limits_{t=1}^T u^t_i(a^t_i)  \right)

No-Regret Learning

In this scenario, we assume that the same nn bidders participate in the same auction for TT rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid. 

We assume that all players choose their bidding stategies aita^t_i to have o(T)o(T) expected regret (no regret), where the regret is defined as 

RT=(maxat=1Tuit(a))(t=1Tuit(ait) )R_T = \left (\max_{a'} \sum\limits_{t = 1}^T u^t_i(a')\right) - \left( \sum\limits_{t=1}^T u^t_i(a^t_i)  \right)

Q: Is there really a way to have o(T)o(T) expected regret surely?

No-Regret Learning

In this scenario, we assume that the same nn bidders participate in the same auction for TT rounds, with their valuations fixed (and kept private). At the end of each round, every bidder knows each other's bid. 

We assume that all players choose their bidding stategies aita^t_i to have o(T)o(T) expected regret (no regret), where the regret is defined as 

RT=(maxat=1Tuit(a))(t=1Tuit(ait) )R_T = \left (\max_{a'} \sum\limits_{t = 1}^T u^t_i(a')\right) - \left( \sum\limits_{t=1}^T u^t_i(a^t_i)  \right)

Q: Is there really a way to have o(T)o(T) expected regret surely?

Yes! For example, the Multiplicative Weights Update algorithm.

For those interested, please refer to Prediction, Learning, and Games.

No-Regret Learning

If every bidder uses an no-regret learning algorithm, we get a vanishing regret sequence a1,a2,a3,...a^1, a^2, a^3, ... of action profiles, That is, for every ii and aiAia_i' \in \cal A_i

limT1Tt=1T(ui(ai,ait;vi)ui(at;vi))0.\operatorname*{lim}_{T\to\infty}\frac{1}{T}\sum_{t=1}^{T}\left(u_{i}(a_{i}^{\prime},{ a}_{-i}^{t};v_{i})-u_{i}({ a}^{t};v_{i})\right)\le0.

No-Regret Learning

If every bidder uses an no-regret learning algorithm, we get a vanishing regret sequence a1,a2,a3,...a^1, a^2, a^3, ... of action profiles, That is, for every ii and aiAia_i' \in \cal A_i

limT1Tt=1T(ui(ai,ait;vi)ui(at;vi))0.\operatorname*{lim}_{T\to\infty}\frac{1}{T}\sum_{t=1}^{T}\left(u_{i}(a_{i}^{\prime},{ a}_{-i}^{t};v_{i})-u_{i}({ a}^{t};v_{i})\right)\le0.

We have the following theorem for PoA of vanishing regret sequence:

If an auction is (λ,μ)(\lambda, \mu)-smooth, then for every valuation profile vv, every vanishing regret sequence of the auction has expected welfare at least λμOPT(v)\frac{\lambda}{\mu} \cdot \mathrm{OPT}(v) as TT \to \infty.

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with unit-demand bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

Critique: How about S1A for more general bidders?

S1A for bidders with Monotone Submodular Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone submodular, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For STS \subseteq T and any item jj, vi(T{j})vi(T)vi(S{j})vi(S) v_i(T \cup \{j\}) - v_i(T)\leq v_i(S \cup \{j\}) - v_i(S)

S1A for bidders with Monotone Submodular Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone submodular, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For STS \subseteq T and any item jj, vi(T{j})vi(T)vi(S{j})vi(S) v_i(T \cup \{j\}) - v_i(T)\leq v_i(S \cup \{j\}) - v_i(S)

diminishing returns

S1A for bidders with Monotone Submodular Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone submodular, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For STS \subseteq T and any item jj, vi(T{j})vi(T)vi(S{j})vi(S) v_i(T \cup \{j\}) - v_i(T)\leq v_i(S \cup \{j\}) - v_i(S)

* Additive and unit-demand valuations are monotone submodular.

S1A for bidders with Monotone Submodular Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone submodular, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For STS \subseteq T and any item jj, vi(T{j})vi(T)vi(S{j})vi(S) v_i(T \cup \{j\}) - v_i(T)\leq v_i(S \cup \{j\}) - v_i(S).

We can prove that S1A for bidders with monotone submodular valuations is (11e,1)(1-\frac{1}{e},1) - smooth, thus the PoA bound is 11e1-\frac{1}{e}.

S1A for bidders with Monotone Submodular Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone submodular, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For STS \subseteq T and any item jj, vi(T{j})vi(T)vi(S{j})vi(S) v_i(T \cup \{j\}) - v_i(T)\leq v_i(S \cup \{j\}) - v_i(S).

We can prove that S1A for bidders with monotone submodular valuations is (11e,1)(1-\frac{1}{e},1) - smooth, thus the PoA bound is 11e1-\frac{1}{e}.

The bound is proven with the composition theorem:

If players have complement-free utility functions, then the simultaneous composition of (λ,μ)(\lambda,\mu)-smooth auctions is again a (λ,μ)(\lambda,\mu)-smooth auction. (See theorem 6.1, Price of Anarchy in Auctions)

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone submodular bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone submodular bidders is at least 11e1-\frac{1}{e}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

Critique: How about S1A for more general bidders?

S1A for bidders with Monotone Subadditive Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone subadditive, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For all S,TS, T, vi(ST)vi(S)+vi(T)v_i(S \cup T) \leq v_i(S) + v_i(T)

* Submodular valuations are subadditive.

S1A for bidders with Monotone Subadditive Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone subadditive, if

1. For STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

2. For all S,TS, T, vi(ST)vi(S)+vi(T)v_i(S \cup T) \leq v_i(S) + v_i(T)

* Submodular valuations are subadditive.

In fact, a PoA bound of 12\frac{1}{2} can be proven for S1A for bidders with monotone subadditive valuations.

* Not proven by the smoothness argument, see Tim's lecture notes.

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone submodular bidders is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone subadditive bidders is at least 12\frac{1}{2}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone submodular bidders is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone subadditive bidders is at least 12\frac{1}{2}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)

Critique: How about S1A for more general bidders?

S1A for bidders with General (Monotone) Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone, if

for all STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

S1A for bidders with General (Monotone) Valuations

For a set II of items, a valuation vi:2IRv_i:2^{I} \to \mathbb R is monotone, if

for all STS \subseteq T, vi(S)vi(T)v_i(S) \leq v_i(T)

Is there a PoA bound for S1A for bidders with general valuations?

Imposssibility Result

With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.

* Even when we consider fixed valuation and MNE.

Imposssibility Result

With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.

The non-deterministic communication complexity of Welfare-Maximization(n) is high

A good PoA bound for simple auctions yields a good communication protocal for Welfare-Maximization(n)

Non-Deterministic Communication Complexity

Fix some function f:X×Y{0,1}f: \cal X \times \cal Y \to \{0,1\}

Alice gets the input xXx \in \cal X, Bob gets the input yYy \in \cal Y, and Charlie gets both xx and yy. Charlie needs to convince Alice and Bob that f(x,y)=1f(x,y) = 1.

Non-Deterministic Communication Complexity

Fix some function f:X×Y{0,1}f: \cal X \times \cal Y \to \{0,1\}

Alice gets the input xXx \in \cal X, Bob gets the input yYy \in \cal Y, and Charlie gets both xx and yy. Charlie needs to convince Alice and Bob that f(x,y)=1f(x,y) = 1.

A protocal for Charlie is a function P:{(x,y)f(x,y)=1}{0,1}P: \{(x,y)|f(x,y) = 1\} \to \{0,1\}^* that "can convince" Alice and Bob.

The complexity of the protocal PP is minf(x,y)=1P(x,y)\min_{f(x,y) = 1} |P(x,y)|.

The non-deterministic communication complexity of ff is the minimum complexity over all protocal.

Non-Deterministic Communication Complexity

Fix some function f:X×Y{0,1}f: \cal X \times \cal Y \to \{0,1\}

Alice gets the input xXx \in \cal X, Bob gets the input yYy \in \cal Y, and Charlie gets both xx and yy. Charlie needs to convince Alice and Bob that f(x,y)=1f(x,y) = 1.

A protocal for Charlie is a function P:{(x,y)f(x,y)=1}{0,1}P: \{(x,y)|f(x,y) = 1\} \to \{0,1\}^* that "can convince" Alice and Bob.

The complexity of the protocal PP is minf(x,y)=1P(x,y)\min_{f(x,y) = 1} |P(x,y)|.

The non-deterministic communication complexity of ff is the minimum complexity over all protocal.

Example: NDISJ(n)

Non-Deterministic Communication Complexity

Fix some function f:X×Y{0,1}f: \cal X \times \cal Y \to \{0,1\}

Alice gets the input xXx \in \cal X, Bob gets the input yYy \in \cal Y, and Charlie gets both xx and yy. Charlie needs to convince Alice and Bob that f(x,y)=1f(x,y) = 1.

A protocal for Charlie is a function P:{(x,y)f(x,y)=1}{0,1}P: \{(x,y)|f(x,y) = 1\} \to \{0,1\}^* that "can convince" Alice and Bob.

The complexity of the protocal PP is minf(x,y)=1P(x,y)\min_{f(x,y) = 1} |P(x,y)|.

The non-deterministic communication complexity of ff is the minimum complexity over all protocal.

There can be more than two parties!

Welfare-Maximization(n)

There are nn players and mm items. Each player has some general monotone valuation for each subset of items.
The goal is to identify the case:

(1) Every allocation of items (partition of items into T1,T2,...,TnT_1, T_2, ..., T_n) has welfare at most 1.

from

(0) There is some allocation with welfare at least nn.

Welfare-Maximization(n)

There are nn players and mm items. Each player has some general monotone valuation for each subset of items.
The goal is to identify the case:

(1) Every allocation of items (partition of items into T1,T2,...,TnT_1, T_2, ..., T_n) has welfare at most 1.

from

(0) There is some allocation with welfare at least nn.

We will prove that the non-deterministic complexity of this "promise" problem is exp(Ω(m/n2))\exp(\Omega(m/n^2)) by reduction!

Multi-Disjointness

There are nn parties, and each party has a string xi{0,1}lx^i \in \{0,1\}^l(viewed as a subset Si[l]S_i \subseteq [l])
The goal is to identify the case:

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset

from

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset

Multi-Disjointness

There are nn parties, and each party has a string xi{0,1}lx^i \in \{0,1\}^l(viewed as a subset Si[l]S_i \subseteq [l])
The goal is to identify the case:

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset

from

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset

Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)\Omega(l/n)

Multi-Disjointness

There are nn parties, and each party has a string xi{0,1}lx^i \in \{0,1\}^l(viewed as a subset Si[l]S_i \subseteq [l])
The goal is to identify the case:

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset

from

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset

Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)\Omega(l/n)

Questions?

Multi-Disjointness

There are nn parties, and each party has a string xi{0,1}lx^i \in \{0,1\}^l(viewed as a subset Si[l]S_i \subseteq [l])
The goal is to identify the case:

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset

from

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset

Fact: The non-determinsitic communication complexity of Multi-Disjointness is Ω(l/n)\Omega(l/n)

Reduction, starts!

Intersecting Family of Partitions

We say tt partitions P1,P2,...,PtP^1,P^2, ..., P^t of mm items into nn players are an "intersecting family", if for all ii,jj,PjiPjii \neq i',j \neq j', P^i_j \cap P^{i'}_{j'} \neq \emptyset

Example: m=6,n=3m = 6, n = 3

P1=[(1,2),(3,4),(5,6)]P^1 = [(1,2), (3,4), (5,6)]

P2=[(3,6),(1,5),(2,4)]P^2 = [(3,6), (1,5), (2,4)]

P3=[(4,5),(2,6),(1,3)]P^3 = [(4,5), (2,6), (1,3)]

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Prove by the probabilistic method!

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Prove by the probabilistic method!

We choose tt (a number to be determined) partitions randomly and independently:

In each partition, each item is assigned to one player, independently and uniformly at random. 

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Prove by the probabilistic method!

We choose tt (a number to be determined) partitions randomly and independently.

For some ii,jji \neq i', j \neq j',

Pr[PjiPji=]=(11n2)mem/n2\Pr[P^i_j \cap P^{i'}_{j'} = \emptyset] = (1-\frac{1}{n^2})^m \leq {e^{-m/n^2}}

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Prove by the probabilistic method!

We choose tt (a number to be determined) partitions randomly and independently.

For some ii,jji \neq i', j \neq j',

Pr[PjiPji=]=(11n2)mem/n2\Pr[P^i_j \cap P^{i'}_{j'} = \emptyset] = (1-\frac{1}{n^2})^m \leq {e^{-m/n^2}}

Union bound:

Pr[ii,jjs.t.PjiPji=] t2n2em/n2\Pr[\exists {i \neq i' , j \neq j'} \mathrm{ s.t. } P^i_j \cap P^{i'}_{j'} = \emptyset]  \leq t^2n^2{e^{-m/n^2}}

Lemma

For every m,n1m,n \geq 1, there exists an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t with t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)).

Prove by the probabilistic method!

Pr[ii,jjs.t.PjiPji=] t2n2em/n2\Pr[\exists {i \neq i' , j \neq j'} \mathrm{ s.t. } P^i_j \cap P^{i'}_{j'} = \emptyset]  \leq t^2n^2{e^{-m/n^2}}

Thus for t<1nem/2n2t < \frac{1}{n} e^{m/2n^2}, the probability (of not being an intersecting family) is less than 1, and the lemma is proven.

Multi-Disjointness to Welfare Maximization

Let S1,S2,...,SnS_1, S_2, ..., S_n be the input to the Multi-Disjointness problem with tt-bit inputs, where t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)) is the same value as in the lemma.

The players coordinate an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t in advance.

Multi-Disjointness to Welfare Maximization

Let S1,S2,...,SnS_1, S_2, ..., S_n be the input to the Multi-Disjointness problem with tt-bit inputs, where t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)) is the same value as in the lemma.

The players coordinate an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t in advance.

When each player ii gets the input SiS_i, one forms the following valuation:

vi(T)={1if TPij for some jSi 0otherwise.v_{i}(T)=\left\{\begin{array}{l l}{{1}}&{{\mathrm{if~}T\supseteq P_{i}^{j}\mathrm{~for~some}\mathrm{~}j\in S_{i}\mathrm{~}}}\\ {{0}}&{{\mathrm{otherwise.}}}\end{array}\right.

Multi-Disjointness to Welfare Maximization

Let S1,S2,...,SnS_1, S_2, ..., S_n be the input to the Multi-Disjointness problem with tt-bit inputs, where t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)) is the same value as in the lemma.

The players coordinate an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t in advance.

When each player ii gets the input SiS_i, one forms the following valuation:

vi(T)={1if TPij for some jSi 0otherwise.v_{i}(T)=\left\{\begin{array}{l l}{{1}}&{{\mathrm{if~}T\supseteq P_{i}^{j}\mathrm{~for~some}\mathrm{~}j\in S_{i}\mathrm{~}}}\\ {{0}}&{{\mathrm{otherwise.}}}\end{array}\right.

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset     \implies the maximum welfare is at most 11

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset     \implies the maximum welfare is at least nn

Multi-Disjointness to Welfare Maximization

Let S1,S2,...,SnS_1, S_2, ..., S_n be the input to the Multi-Disjointness problem with tt-bit inputs, where t=exp(Ω(m/n2))t = \exp(\Omega(m/n^2)) is the same value as in the lemma.

The players coordinate an intersecting family P1,P2,...,PtP^1, P^2, ..., P^t in advance.

When each player ii gets the input SiS_i, one forms the following valuation:

vi(T)={1if TPij for some jSi 0otherwise.v_{i}(T)=\left\{\begin{array}{l l}{{1}}&{{\mathrm{if~}T\supseteq P_{i}^{j}\mathrm{~for~some}\mathrm{~}j\in S_{i}\mathrm{~}}}\\ {{0}}&{{\mathrm{otherwise.}}}\end{array}\right.

(1) For all i,ii,i', SiSi=S^i \cap S^{i'}= \emptyset     \implies the maximum welfare is at most 11

(0) i=1nSi\large\cap_{i=1}^n S_i \neq \emptyset     \implies the maximum welfare is at least nn

Reduction complete!

Since we have the Ω(l/n)\Omega(l/n) bound for Multi-Disjointness, the non-deterministic communication complexity of Welfare-Maximization(n) is Ω(t/n)=exp(Ω(m/n2))\Omega(t/n) = \exp(\Omega(m/n^2)).

Imposssibility Result

With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible.

The non-deterministic communication complexity of Welfare-Maximization(n) is exp(Ω(m/n2))\exp({\Omega(m/n^2)})

A good PoA bound for simple auctions yields a good communication protocal for Welfare-Maximization(n)

?

Theorem

Fix a class V\cal V of possible bidder valuations.

Suppose that, for some α(0,1]\alpha \in (0,1], there is no nondeterministic communication protocal with subexponential (in mm) communication for the following problem:

(1) Every allocation has welfare at most αW\alpha \cdot W^*

(0) There is some allocation with welfare at least WW^*

Let ϵ1/poly(m,n)\epsilon \geq 1/\mathrm{poly}(m,n).  Then, for every auction with sub-doubly-exponential (in mm) strategies per player, the worst-case PoA of ϵ\epsilon-approximate MNE with bidder valuations in V\cal V is at most α\alpha.

Theorem

Fix a class V\cal V of possible bidder valuations.

Suppose that, for some α(0,1]\alpha \in (0,1], there is no nondeterministic communication protocal with subexponential (in mm) communication for the following problem:

(1) Every allocation has welfare at most αW\alpha \cdot W^*

(0) There is some allocation with welfare at least WW^*

Let ϵ1/poly(m,n)\epsilon \geq 1/\mathrm{poly}(m,n).  Then, for every auction with sub-doubly-exponential (in mm) strategies per player, the worst-case PoA of ϵ\epsilon-approximate MNE with bidder valuations in V\cal V is at most α\alpha.

Theorem

Fix a class V\cal V of possible bidder valuations.

Suppose that, for some α(0,1]\alpha \in (0,1], there is no nondeterministic communication protocal with subexponential (in mm) communication for the following problem:

(1) Every allocation has welfare at most αW\alpha \cdot W^*

(0) There is some allocation with welfare at least WW^*

Let ϵ1/poly(m,n)\epsilon \geq 1/\mathrm{poly}(m,n).  Then, for every auction with sub-doubly-exponential (in mm) strategies per player, the worst-case PoA of ϵ\epsilon-approximate MNE with bidder valuations in V\cal V is at most α\alpha.

Theorem

Fix V\cal V to be general monotone valuations.

Suppose that, for some α(0,1]\alpha \in (0,1], there is no nondeterministic communication protocal with subexponential (in mm) communication for the following problem:

(1) Every allocation has welfare at most αW\alpha \cdot W^*

(0) There is some allocation with welfare at least WW^*

Let ϵ1/poly(m,n)\epsilon \geq 1/\mathrm{poly}(m,n).  Then, for every auction with sub-doubly-exponential (in mm) strategies per player, the worst-case PoA of ϵ\epsilon-approximate MNE with bidder valuations in V\cal V is at most α\alpha.

Theorem

Fix V\cal V to be general monotone valuations.

Suppose that, for some α(0,1]\alpha \in (0,1], there is no nondeterministic communication protocal with subexponential (in mm) communication for the following problem:

(1) Every allocation has welfare at most αW\alpha \cdot W^*

(0) There is some allocation with welfare at least WW^*

Let ϵ1/poly(m,n)\epsilon \geq 1/\mathrm{poly}(m,n).  Then, for every auction with sub-doubly-exponential (in mm) strategies per player, the worst-case PoA of ϵ\epsilon-approximate MNE with bidder valuations in V\cal V is at most α\alpha.

With general valuations, every simple auction can have equilibria with social welfare arbitrarily worse than the maximum possible!

Lemma

For every ϵ>0\epsilon > 0 and every game with nn players with strategy sets A1,A2,...,AnA_1, A_2, ... , A_n , there exists an ϵ\epsilon-approximate MNE with description length polynomial in nn, log(maxAi)\log(\max |A_i|), and 1/ϵ1/\epsilon .

Lemma

For every ϵ>0\epsilon > 0 and every game with nn players with strategy sets A1,A2,...,AnA_1, A_2, ... , A_n , there exists an ϵ\epsilon-approximate MNE with description length polynomial in nn, log(maxAi)\log(\max |A_i|), and 1/ϵ1/\epsilon .

Proof of Theorem:

So if  maxAi\max|A_i| is sub-doubly-exponential, we can write down the ϵ\epsilon-approximate MNE, and the contribution to the welfare of each player under the ϵ\epsilon-approximate MNE in subexponential bits.
If PoA >α> \alpha, this is in fact a proof for "every allocation has welfare at most αW\alpha \cdot W^*"(from "There is some allocation with welfare at least
 WW^*").

Proof of Theorem:

So if  maxAi\max|A_i| is sub-doubly-exponential, we can write down the ϵ\epsilon-approximate MNE, and the contribution to the welfare of each player under the ϵ\epsilon-approximate MNE in subexponential bits..
If PoA >α> \alpha, this is in fact a proof for "every allocation has welfare at most αW\alpha \cdot W^*"(from "There is some allocation with welfare at least
 WW^*").

The players only need to check:

1. The declared ϵ\epsilon-approximate MNE is "true" for him/her.

2. The declared contribution to the welfare of his/hers is "true"

3. The sum of all declared players' contribution to the welfare is smaller than αW\alpha \cdot W^*

Proof of Theorem:

So if  maxAi\max|A_i| is sub-doubly-exponential, we can write down the ϵ\epsilon-approximate MNE, and the contribution to the welfare of each player under the ϵ\epsilon-approximate MNE in subexponential bits..
If PoA >α> \alpha, this is in fact a proof for "every allocation has welfare at most αW\alpha \cdot W^*"(from "There is some allocation with welfare at least
 WW^*").

The players only need to check:

1. The declared ϵ\epsilon-approximate MNE is "true" for him/her.

2. The declared contribution to the welfare of his/hers is "true"

3. The sum of all declared players' contribution to the welfare is smaller than αW\alpha \cdot W^*

Why is this a valid proof to the players?

Proof of Theorem:

So if  maxAi\max|A_i| is sub-doubly-exponential, we can write down the ϵ\epsilon-approximate MNE, and the contribution to the welfare of each player under the ϵ\epsilon-approximate MNE in subexponential bits..
If PoA >α> \alpha, this is in fact a proof for "every allocation has welfare at most αW\alpha \cdot W^*"(from "There is some allocation with welfare at least
 WW^*").

The players only need to check:

1. The declared ϵ\epsilon-approximate MNE is "true" for him/her.

2. The declared contribution to the welfare of his/hers is "true"

3. The sum of all declared players' contribution to the welfare is smaller than αW\alpha \cdot W^*

But we assumed there is no subexponential communication protocal!     PoAα\implies \mathrm{PoA} \leq \alpha

What we've done

  • Prove that the PoA of single-item first-price auction is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone submodular bidders is at least 11e1-\frac{1}{e}
  • Prove that the PoA of S1A with monotone subadditive bidders is at least 12\frac{1}{2}
  • Prove PoA bounds for general smooth auctions.(Also in regret-vanishing sequences.)
  • Prove that no PoA bounds holds for "simple auctions" for general monotone valuations (considering approximate equilibrium).

Open Problems

  • Can we prove that no PoA bound holds for general monotone valuations considering exact equilibria?
  • What is the PoA of first-price single-item auctions? (Currently known in [11e,0.87][1-\frac{1}{e},0.87])
  • PoA in reverse auctions?

AGT beyond PoA of Auctions

  • Complexity of computing a Nash Equilibrium (PLS, PPAD)
  • Algorithmic Mechanism Design (Knapsack auction)
  • ...

End!

Kolmogorov Complexity and the Incompressibility Method

Eric Xiao 2024/8/22

What's the next number?

1,2,4,8,16,1,2,4,8,16,

What's the next number?

1,2,4,8,16,1,2,4,8,16,

71227122?

3232?

What's the next number?

1,2,4,8,16,1,2,4,8,16,

71227122?

3232?

"less random"?

What is randomness?

What is randomness?

Random objects are harder to describe/compress?

1,2,4,8,16,71221,2,4,8,16,7122

1,2,4,8,16,321,2,4,8,16,32

A sequence aa of length 66 where a1=1a_1 = 1 and ai+1=2aia_{i+1} = 2a_i.

The sequence 1, 2, 4, 8, 16, 7122.

A sequence aa of length 66 where a1=1a_1 = 1 and ai+1=2aia_{i+1} = 2a_i but a6=7122a_6 = 7122.

1,2,4,8,16,71221,2,4,8,16,7122

1,2,4,8,16,321,2,4,8,16,32

k = 1
for i in range(5):
  print(k)
  k *= 2
k = 1
for i in range(4):
  print(k)
  k *= 2
print(7122)
A = [1,2,4,8,16,7122]
for i in range(5):
  print(A[i])

Kolmogorov Complexity:

How short can a string be compressed?

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

Kolmogorov Complexity

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

A more formal definition

Cf(x)=min{l(p):f(p)=x}C_f(x) = \min\{l(p):f(p) = x\}

Where ff is some (partial) function.

Kolmogorov Complexity

C(x)=min{l(p):p is an encoding/program of x}C(x) = \min\{l(p):p \text{ is an encoding/program of }x\}

Think of x,px,p as binary strings(or natural numbers).

A more formal definition

Cf(x)=min{l(p):f(p)=x}C_f(x) = \min\{l(p):f(p) = x\}

Where ff is some (partial) function.

How should we choose ff? (English, python, Eric Xiao, ...?)

Kolmogorov Complexity

Maybe ff should be the "minimal" (additively optimal) function:

g,x,Cf(x)Cg(x)+cg\forall g,x, C_f(x) \leq C_g(x) + c_g

Kolmogorov Complexity

Such ff doesn't exist:

Let g(i)=xig(i) = x_i, where Cf(xi)iC_f(x_i) \geq i

Maybe ff should be the "minimal" (additively optimal) function:

g,x,Cf(x)Cg(x)+cg\forall g,x, C_f(x) \leq C_g(x) + c_g

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Let f0f_0 be a function computed by a universal TM (T0T_0) with input  11..110np11..110np (with l(n)l(n) ones ) which simulates TnT_n on input pp

Kolmogorov Complexity

Perhaps we should not consider all functions.

Computable functions?

Let f0f_0 be a function computed by a universal TM (T0T_0) with input  11..110np11..110np (with l(n)l(n) ones ) which simulates TnT_n on input pp

We have Cf0(x)Cfn(x)+2l(n)+1C_{f_0}(x) \leq C_{f_n}(x) + 2l(n)+1

(since f0(11..10np)=fn(p)f_0(11..10np) = f_n(p))

f0f_0 is additively optimal!

?

Kolmogorov Complexity

C(x)=Cf0(x)=min{l(p):f0(p)=x}C(x) = C_{f_0}(x) = \min\{l(p):f_0(p) = x\}

Kolmogorov Complexity

C(x)=Cf0(x)=min{l(p):f0(p)=x}C(x) = C_{f_0}(x) = \min\{l(p):f_0(p) = x\}

We can view f0f_0 as a special compiler.

Is the compiler of C/Java additively optimal?

Conditional Kolmogorov Complexity

C(xy)=Cf0(xy)=min{l(p):f0(y,p)=x}C(x|y) = C_{f_0}(x|y) = \min\{l(p):f_0(\langle y,p\rangle) = x\}

C(x)=C(xϵ)C(x) = C(x|\epsilon)

We also have f,x,y\forall f,x,y

Cf0(xy)Cf(xy)+cfC_{f_0}(x|y) \leq C_f(x|y) + c_f

Upper bounds

C(x)l(x)+cC(x) \leq l(x) + c

C(xy)C(x)+cC(x|y) \leq C(x)+c

Uncomputability

C(x)C(x) is uncomputable

Uncomputability

C(x)C(x) is uncomputable

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

Uncomputability

C(x)C(x) is uncomputable

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

Berry's Paradox:"The smallest positive integer not definable in under sixty letters."

Uncomputability

C(x)C(x) is uncomputable on any infinite set of points

function GenerateComplexString()
    for i = 1 to infinity:
        for each string s of length exactly i
            if KolmogorovComplexity(s) ≥ 8000000000
                return s

(Assume the length of KolmogorovComplexity() is 700000000700000000)

?

Incompressibility Theorem \infty \star

Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Incompressibility Theorem \infty \star

Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

When c=0c=0?  c=1?c=1?

Incompressibility Theorem \infty \star

Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Because there are only i=0logmc12i=m2c1\sum\limits_{i=0}^{\log m-c-1} 2^i= m2^{-c}-1 strings with length less than logmc\log m-c.

Example - Infinite Primes

Assume that there are only kk primes. Then for each nNn \in \mathbb N, we can write n=i=1kpiαin = \prod\limits_{i=1}^k p_i^{\alpha_i}. Thus we can describe every integer in [1,t][1,t] by kloglogtk \log\log t bits......

Incompressibility theorem: Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

Example - Infinite Primes

Assume that there are only kk primes. Then for each nNn \in \mathbb N, we can write n=i=1kpiαin = \prod\limits_{i=1}^k p_i^{\alpha_i}. Thus we can describe every integer in [1,t][1,t] by kloglogtk \log\log t bits......

Incompressibility theorem: Let cNc \in \mathbb N and AA be a set of size mm,

then there are at least m(12c)+1m(1-2^{-c})+1 elements xAx\in A such that

C(xy)logmcC(x|y) \geq \log m -c, for any fixed yy.

π(n)Ω(lognloglogn)\pi(n) \in \Omega(\frac{\log n}{\log \log n})

Example - TM complexity

Theorem: Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Example - TM complexity

Theorem: Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

eg. RADAR

Example - TM complexity

Theorem: Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Assume a TM TT:

  • each state can be described by \ell bits
  • always terminates with the reading head at the rightmost position of the input
  • computes PALINDROME

Example - TM complexity

Theorem: Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Assume a TM TT:

  • each state can be described by \ell bits
  • always terminates with the reading head at the rightmost position of the input
  • computes PALINDROME

Define the crossing sequence of a grid gg:   the sequence of states of the TM TT after the reading head crosses the right boundary of gg

Example - TM complexity

By the incompressibility theorem, we can find length-nn string xx with C(xT,n)nC(x|T,n) \geq n.

*If the runtime of the TM TT is less than n2\frac{n^2}{\ell} for all strings of length 4n4n, then we can describe every string of length nn with n2+O(logn)\frac{n}{2}+O(\log n) bits.

Theorem: Checking if a string is palindrome needs Ω(n2)\Omega(n^2) time for any one-tape Turing Machine.

Example - TM complexity

Consider running TT on x02nxˉx 0^{2n} \bar x, if the runtime is less than n2\frac{n^2}{\ell}, then there exists a "middle" grid g0g_0 with crossing sequence shorter than n2\frac{n}{2 \ell}.

In fact, given T,nT,n, we can recover xx by the position of g0g_0(O(logn)O(\log n) bits) and the crossing sequence of g0g_0(n2\frac{n}{2} bits).

Thus C(xT,n)n2+O(logn)C(x|T,n) \leq \frac{n}{2} + O(\log n)

*If the runtime of the TM TT is less than n2\frac{n^2}{\ell} for all string of length 4n4n, then we can describe every string of length nn with n2+O(logn)\frac{n}{2}+O(\log n) bits.

Incompressibility Method

If the statement doesn't hold, then there will be too many effectively compressible objects.

?

More Examples: Combinatorial Properties

Some Useful Bounds

nnen1n!nn+1en1\dfrac {n^n} {e^{n - 1} } \le n! \le \dfrac {n^{n + 1} } {e^{n - 1} }

(nk)kCknnkk!<(nek)k\left({\frac{n}{k}}\right)^{k}\leq C^n_k\leq{\frac{n^{k}}{k!}}\lt \left({\frac{n e}{k}}\right)^{k}

Transitive Tournaments

Tournament: complete directed graph

 

 

 

 

Transitive tournament: acyclic tournament

 

 

 

 

v(n):v(n): Largest integer such that every tournament on nn nodes contains a transitive subtournament on v(n)v(n) nodes.

1

2

4

3

Transitive Tournament

Tournament: complete directed graph

 

Transitive tournament: acyclic tournament

 

v(n):v(n): Largest integer such that every tournament on nn nodes contains a transitive subtournament on v(n)v(n) nodes.

logn\lfloor \log n \rfloor \leq v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

Counting

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

v:=2+2lognv:= 2 + \lfloor 2\log n \rfloor

A:A: subset of nn vertices of size vv

σ:\sigma: permutation on [v][v]

ΓA,σ:\Gamma_{A,\sigma}: the set of nn- vertex tournaments where AA is a subtournament of it with "order" σ\sigma

We count the number d(Γ)d(\Gamma') of nn- vertex tournaments with a transitive subtournament of size vv

d(Γ)Aσd(ΓA,σ)=Cvnv!2C2nC2v<2C2nd(\Gamma') \leq \sum_A \sum_\sigma d(\Gamma_{A,\sigma}) = C^n_v v! 2^{C^n_2-C^v_2} < 2^{C^n_2}

(Cvnv!<nv2C2v) (C^n_v v! < n^v \leq 2^{C^v_2})

Probabilistic Method

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

v:=2+2lognv:= 2 + \lfloor 2\log n \rfloor

A:A: subset of nn vertices of size vv

σ:\sigma: permutation on [v][v]

Let TT be the random variable uniformly distributed over all nn-vertex tournaments.

We calculate the probability P(TΓ)P(T\in \Gamma').

P(TΓ)AσP(TΓA,σ)=Cvnv!2C2v<1P(T\in \Gamma') \leq \sum_A \sum_\sigma P(T \in \Gamma_{A,\sigma}) = C^n_v v! 2^{-C^v_2} < 1

(Cvnv!<nv2C2v) (C^n_v v! < n^v \leq 2^{C^v_2})

Incompressibility Method

v(n)1+2lognv(n) \leq 1 + \lfloor 2\log n \rfloor

If v(n)v(n) is too large, then we can effectively compress every nn-vertex tournament.

However, by the incompressibility method, there is some tournament TT', where C(Tn,p)n(n1)/2C(T'|n,p) \geq n(n-1)/2.

(pp is some decoding program)

For every tournament TT of size nn, we can describe it by v(n)logn+n(n1)/2v(n)(v(n)1)/2v(n) \lfloor \log n \rfloor+ n(n-1)/2 - v(n)(v(n)-1)/2 bits.

Thus v(n)lognv(n)(v(n)1)/20v(n) \lfloor \log n \rfloor - v(n)(v(n)-1)/2 \geq 0, and

v(n)2logn+1v(n) \leq 2 \lfloor \log n\rfloor + 1.

?

Last Example:
Regular Languages

Regular Language

a language accepted by some DFA

a language decided by some constant-memory one-pass algorithm.

    \iff

Regular Language

a language accepted by some DFA

a language decided by some constant-memory one-pass algorithm.

    \iff

e.g. L={01}={1,01,001,0001,...}L = \{0^*1\}=\{1,01,001,0001,...\}

How to prove that a language is regular?

How to prove that a language is regular?

Construct a DFA/constant-memory one-pass algorithm for it.

How to prove that a language is regular?

Construct a DFA/constant-memory one-pass algorithm for it.

How to prove that a language is non-regular?

Pumping Lemma

Pumping Lemma

Might not be easy to use.

Not characterizing.

Example: Another possible way

Is L={0k1kk1}L = \{0^k 1^k | k \geq 1\} a regular language?

Example: Another possible way

Is L={0k1kk1}L = \{0^k 1^k | k \geq 1\} a regular language?

Suppose LL is accepted by some DFA AA. Then we can recover kk from AA and qq, the state of AA after processing 0k0^k. Thus for all k1k\geq 1, C(k)=O(1)C(k) = O(1).

KC-Regularity

Suppose LL is regular, and Lx:={yxyL}L_x := \{y|xy \in L \}. If yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

KC-Regularity

Suppose LL is regular, and Lx:={yxyL}L_x := \{y|xy \in L \}. If yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

*What does "the nn-th string" mean?

More Examples

Is L:={1pp is prime}L := \{1^p |p\text{ is prime}\} a regular language?

Is L:={xxˉ}L := \{x \bar x \} a regular language?

Regular KC-Characterization

In fact, we have

LL is regular

For all x,nx,n, if yy is the nn-th string in LxL_x, then C(y)C(n)+O(1)C(y) \leq C(n) + O(1).

\Leftrightarrow

Thus, (in theory), we can also prove regularity!

Takeaways?

1. What is Kolmogorov complexity?

The Kolmogorov complexuty of a string (or number or sequence) is the minimum number of bits it can be compressed to.

2. What is the incompressibility method?

If an object satisfies some property, it can be efficiently compressed, but not everything can be efficiently compressed.

Some Questions

1. Alice and Bob are playing the number guessing game. Alice chooses some x[n]x \in [n], and Bob can ask her yes/no questions. If Alice can lie once, what is the minimum number cc where Bob can always guess the right answer after asking cc questions?

4. Why is computability important?

5. Ask a question!

3. Prove any (nontrivial?) statement you like via the incompressibility method.

2. What are the differences between counting, the probabilistic method, and the incompressibility method?

References

[1] Wikipedia contributors. (2023, March 5). Incompressibility method. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Incompressibility_method&oldid=1143081380

[2] Wikipedia contributors. (2023, August 22). Kolmogorov complexity. In Wikipedia, The Free Encyclopedia. Retrieved 15:03, August 27, 2023, from https://en.wikipedia.org/w/index.php?title=Kolmogorov_complexity&oldid=1171617849

[3] Ming Li , Paul Vitányi. An Introduction to Kolmogorov Complexity and Its Applications, 2008. 

Made with Slides.com