How NP got a new definition: Probabilistically Checkable Proofs NP-hard problems

Download Report

Transcript How NP got a new definition: Probabilistically Checkable Proofs NP-hard problems

How NP got a new definition:
Probabilistically Checkable Proofs
(PCPs) & Approximation Properties of
NP-hard problems
SANJEEV ARORA
PRINCETON UNIVERSITY
Talk Overview
• Recap of NP-completeness and its philosophical
importance.
• Definition of approximation.
• How to prove approximation is NP-complete
(new definition of NP; PCP Theorem)
• Survey of approximation algorithms.
A central theme in modern TCS:
Computational Complexity
How much time (i.e., # of basic operations) are needed
to solve an instance of the problem?
Example: Traveling Salesperson
Problem on n cities
n =49
Number of all possible salesman tours = n!
(> # of atoms in the universe for n =49)
One key distinction: Polynomial time (n3, n7 etc.)
versus
Exponential time (2n, n!, etc.)
Is there an inherent difference
between
being creative / brilliant
• Writing the Moonlight Sonata
• Proving Fermat’s Last Theorem
• Coming up with a low-cost
salesman tour
and
being able to appreciate
creativity / brilliance?
• Appreciating/verifying
any of the above
When formulated as “computational effort”, just the P vs NP
Question.
P vs. NP
NPC
NP
P
“YES” answer has
certificate of O(nc) size,
verifiable in O(nc’) time.
Solvable in O(nc) time.
NP-complete: Every NP problem is reducible to it in
O(nc) time. (“Hardest”)
e.g., 3SAT: Decide satisfiability of a boolean formula like
Practical Importance of P vs NP: 1000s of optimization
problems are NP-complete/NP-hard. (Traveling Salesman,
CLIQUE, COLORING, Scheduling, etc.)
Pragmatic Researcher
“Why the fuss? I am perfectly content with approximately
optimal solutions.” (e.g., cost within 10% of optimum)
Good news: Possible for quite a few problems.
Bad News: NP-hard for many problems.
Approximation Algorithms
MAX-3SAT: Given 3-CNF formula , find assignment
maximizing the number of satisfied clauses.
An -approximation algorithm is one that for every
formula, produces in polynomial time an assignment that
satisfies at least OPT/ clauses. ( ¸ 1).
Good News: [KZ’97] An 8/7-approximation algorithm exists.
Bad News: [Hastad’97] If P  NP then for every  > 0, an
(8/7 -)-approximation algorithm does not exist.
Observation (1960’s thru ~1990)
NP-hard problems differ with respect to approximability
[Johnson’74]: Provide explanation? Classification?
Last 15 years: Avalanche of Good and Bad news
Next few slides: How to rule out existence
of good approximation algorithms
(New definition of NP via PCP Theorems
and why it was needed)
Recall: “Reduction”
“If you give me a place to
stand, I will move the earth.”
– Archimedes (~ 250BC)
a 1.01-approximation for MAX-3SAT
“If you give me a polynomial-time algorithm
for 3SAT, I will give you a polynomial-time
algorithm for every NP problem.”
--- Cook, Levin (1971)
[A., Safra] [A., Lund, Motwani, Sudan, Szegedy] 1992
“Every instance of an NP problem can be disguised as
an instance of 3SAT.” MAX-3SAT
Desired
Way to disguise instances of any NP problem as
instances of MAX-3SAT s.t.
“Yes” instances turn into satisfiable formulae
“No” instances turn into formulae in which < 0.99
fraction of clauses can be simultaneously satisfied
“Gap”
Cook-Levin reduction doesn’t produce instances
where approximation is hard.
Transcript of computation
?
Main point: Express
these as boolean formula
Transcript is “correct” if it satisfies all “local” constraints.
But, there always exists a transcript that satisfies almost all
local constraints! (No “Gap”)
New definition of NP….
Recall: Usual definition of NP
nc
n
INPUT x
CERTIFICATE

M
x is a “YES” input
 there is a  s.t. M accepts (x, )
x is a “NO” input
 M rejects (x, ) for every 
NP = PCP (log n, 1)
[AS’92][ALMSS’92]; inspired by [BFL’90], [BFLS’91][FGLSS’91]
nc
n
INPUT x
Uses O(log n)
random bits
CERTIFICATE
M

Reads Fixed number of bits
(chosen in randomized fashion)
(Only 3 bits ! (Hastad 97))
Many other
x is a “YES” input
“PCP Theorems”
 there is a  s.t. Pr [M accepts (x, ) ] = 1
known now.
x is a “NO” input
for every ,
Pr [ M rejects (x, ) ] > 1/2
Disguising an NP problem as MAX-3SAT
?
INPUT x
O(lg n) random bits
M
Note: 2O(lg n) = nO(1).
) M ≡ nO(1) constraints, each on O(1) bits
x is YES instance
) All are satisfiable
x is NO instance
) · ½ fraction satisfiable
“gap”
Of related interest….
Do you need to read a math proof completely to
check it?
Recall: Math can be axiomatized (e.g., Peano Arithmetic)
Proof = Formal sequence of derivations from axioms
PCP Theorem
Verification of math proofs
(spot-checking)
n bits
Theorem
Proof
M
O(1) bits
M runs in poly(n) time
•Theorem correct  there is a proof that M accepts w. prob. 1
•Theorem incorrect  M rejects every claimed proof w. prob 1/2
Known Inapproximability Results
The tree of reductions [AL ‘96]
MAX-3SAT [FGLSS ’91,
MAX-3SAT(3)
Class I
1+
Metric TSP
Vertex Cover
MAX-CUT
STEINER...
[LY ’93,
ABSS ’93]
LABEL COVER
SET COVER
Class II
O(lg n)
HITTING SET
DOMINATING SET
HYPERGRAPH TRAVERSAL ...
Class III
1-
2(lg n)
NEAREST VECTOR
MIN-UNSATISFY
QUADRATIC PROGRAMMING
LONGEST PATH
...
BS ‘89]
CLIQUE
COLORING
Class IV
n
INDEPENDENT SET
BICLIQUE COVER
FRACTIONAL
COLORING MAXPLANAR SUBGRAPH
MAX-SET PACKING
MAX-SATISFY
Proof of PCP Theorems
( Some ideas )
Need for “robust” representation
:
x x
x
1 0 0 1 0 1 1 1 0 1 0 0 0 1 0 1
O(lg n) random bits
3 bits
Randomly corrupt 1% of 
Correct proof still accepted with 0.97- probability!
Original proof of PCP Thm used polynomial representations,
Local “testing” algorithms for polynomials, etc. (~30-40 pages)
New Proof (Dinur’06); ~15-20 pages
Repeated applications of two operations on the clauses:
Globalize: Create new constraints using “walks” in the
adjacency graph of the old constraints.
Domain reduction: Change constraints so variables take values
in a smaller domain (e.g., 0,1)
(uses ideas from old proof)
Unique game conjecture and why it is useful
Problem: Given system of equations modulo p (p is prime).
7x2 + 2x4 = 6
5x1 + 3x5 = 2

2 variables per equation

7x5 + x2 = 21
UGC (Khot03): Computationally intractable to distinguish between the cases:
• 0.99 fraction of equations are simultaneously satisfiable
• no more than 0.001 fraction of equations are simultaneously satisfiable.
Implies hardness of approximating vertex cover, max-cut, etc.
(K04), (KR05)(KKMO05)
Applications of PCP Techniques: Tour
d’Horizon
• Locally checkable / decodable codes.
• List decoding of error-correcting codes.
• Private Info Retrieval
[Katz, Trevisan 2000]
• Zero Knowledge arguments / CS proofs
•
[Sudan ’96, Guruswami-Sudan]
[Kilian ‘94] [Micali]
[Lipton ‘88]
Amplification of hardness / derandomization
[A., Sudan ’97]
[Sudan, Trevisan, Vadhan]
• Constructions of Extractors.
[Safra, Ta-shma, Zuckermann]
• Property testing
[Goldreich, Goldwasser,
Ron ‘97]
[Shaltiel, Umans]
Approximation algorithms: Some major
ideas
How can you prove that the solution you found has
cost at most 1.5 times (say) the optimum cost?
• Relax, solve, and round : Represent problem using a
linear or semidefinite program, solve to get fractional
solution, and round to get an integer solution. (e.g.,
MAX-CUT, MAX-3SAT, SPARSEST CUT)
• Primal-dual: “Grow” a solution edge by edge; prove its
near optimality using LP duality. (Usually gives faster
algorithms.) e.g., NETWORK DESIGN, SET COVER
• Show existence of “easy to find” near-optimal solutions:
e.g., Euclidean TSP and Steiner Tree
Next few slides: The semidefinite programming approach
What is semidefinite programming?
Ans. Generalization of linear programming; variables are
vectors instead of fractions. “Nonlinear optimization.”
[Groetschel, Lovasz, Schrijver ’81]; first used in approximation algorithms
by [Goemans-Williamson’94]
Main Idea:
v2
v1
G = (V,E)
vn
n vertices
Rn
v3
n vectors, d(vi,vj)
satisfy some constraints.
“Round”
Ex: 1.13 ratio for MAX-CUT, MAX-2SAT [GW ’93]
How do you analyze these vector
O(lg n) ratio for min-multicut, sparsest cut. [LLR ’94, AR ’94]
programs?
n1/4-coloring of 3-colorable graphs. [KMS ’94]
(lg n)O(1) ratio for min-bandwidth and related problems [F ’98, BKRV ’98]
8/7 ratio for MAX-3SAT [KZ ’97]
Ans. Geometric arguments; sometimes very complicated
plog n-approximation for graph partitioning problems (ARV04)
Ratio 1.13.. for MAX-CUT
G = (V,E) Find
that maximizes capacity
Quadratic Programming Formulation
Semidefinite Relaxation [DP ’91, GW ’93]
.
[GW ’93]
Rn
v6
v5
Randomized Rounding
v2
v1
v3
[GW ’93]
Form a cut by partitioning v1,v2,...,vn
around a random hyperplane.
SDPOPT
vi
ij
vj
Old math rides to the rescue...
sparsest cut: edge expansion
Input: A graph G=(V,E).
For a cut (S,S) let E(S,S) denote the edges
crossing the cut.
S
The sparsity of S is the value
The SPARSEST CUT problem is to find the cut which minimizes (S).
SDPs used to give plog n -approximation involves proving a
nontrivial fact about high-dimensional geometry [ARV04]
ARV structure theorem
Arora, Rao, and Vazirani showed how the SDP could be rounded
to obtain an
approximation to Sparsest Cut (2004)
After we have such A and B, it is easy to
ARV structure theorem:
extend them to a good sparse cut in G.
If the points xu 2 Rn are well-spread, e.g.
u,v (xu-xv)2 ¸ 0.1 and xu2 · 10 for u 2 V
and d(u,v) = (xu-xv)2 is a metric, then…
There exist two large, well-separated sets
A, B µ {x1, x2, …, xn}
with |A|,|B| ¸ 0.1 n and
A
B
Unexpected progress in
other disciplines…
ARV structure theorem led to new understanding of
the interrelationship between l1 and l2 norms
(resolved open question in math)
l1 distances among n points can be realized as
l2 distances among some other set of n points, and
the distortion incurred is only plog n
[A., Lee, Naor’05], building upon [Chawla Gupta Raecke’05]
Theory of Approximability?
Desired Ingredients:
1. Definition of approximation-preserving reduction.
2. Use reductions and algorithms to show:
factor ()
Approx.
upto ()
All interesting
problems
Partial Progress
Max-SNP:
Problems similar to MAX-3SAT.
[PY ’88]
RMAX(2):
Problems similar to CLIQUE.
[PR ‘90]
F+2(1):
Problems similar to SET COVER.
[KT ’91]]
MAX-ONES CSP, MIN-CSP,etc. (KST97, KSM96)
Further Directions
1.
Investigate alternatives to approximation
•
•
2.
Average case analysis
Slightly subexponential algorithms (e.g. 2o(n) algorithm for
CLIQUE??)
Resolve the approximability of graph partitioning
problems. (BISECTION, SPARSEST CUT,
plog n vs loglog n??) and Graph Coloring
3. Complete the classification of problems w.r.t.
approximability.
4. Simplify proofs of PCP Thms even further.
5. Resolve “unique games”conjecture.
6. Fast solutions to SDPs? Limitations of SDPs?
Attributions
[Fortnow, Rompel, Sipser ’88]
Definition
of PCP
[Feige, Goldwasser, Lovász,
Safra, Szegedy ’91]
[Arora, Safra ’92]
Polynomial
Encoding
Method
Verifier
Composition
[Arora, Safra ’92]
[Lund, Fortnow, Karloff, Nisan ’90]
[Shamir ’90]
[Babai, Fortnow ’90]
[Babai, Fortnow, Levin, Szegedy ’91]
PCP
Hardness
of Approx.
[FGLSS ’91]
[ALMSS ’92]
Fourier
Transform
Technique
[Håstad ’96, ’97]
Constraint Satisfaction Problems
[Schaefer ’78]
Let F = a finite family of boolean constraints.
An instance of CSP(F):
x1
x2 . . . . . . . . . . . . xn
g1
variables
g2 . . . . . . . . . . . . gm
functions
from F
Ex:
Dichotomy Thm:
Iff F is 0-valid, 1-valid,
weakly positive or
negative, affine, or
2CNF
{CSP(F) : F is finite}
P
NP Complete
[Creignou ‘96]
MAX-CSP
MAX-SNP-hard
(1+) ratio is
NP-hard
P
Iff F is 0-valid,
1-valid, or
2-monotone
MAX-ONES-CSP
[KSW ‘97]
[Khanna,
Sudan,
Williamson ‘97]
(Supercedes
MAXSNP work)
Ex:
P
1+
n
MIN-ONES-CSP
Feasibilty
NP-hard
Feasibility is
undecidable
[KST ‘97]
Ex:
P
n
Feasibilty
NP-hard
MIN-HORN-DELETION-complete
1+
NEAREST-CODEWORD-complete
Geometric Embeddings of Graphs
v2
G = (V,E)
n vertices
Rn
v1
vn
v3
n vectors, d(vi,vj)
satisfy some constraints.
Ex: 1.13 ratio for MAX-CUT, MAX-2SAT [GW ’93]
O(lg n) ratio for min-multicut, sparsest cut. [LLR ’94, AR ’94]
n1/4-coloring of 3-colorable graphs. [KMS ’94]
(lg n)O(1) ratio for min-bandwidth and related problems [F ’98, BKRV ’98]
8/7 ratio for MAX-3SAT [KZ ’97]
plog n-approximation for graph partitioning problems (ARV04)
Example: Low Degree Test
F =GF(q)
Does
Is ff agree
a
with a
degree d polynomial ?
f:F
m
!F
in 90% of the points?
Easy: f isIffaon
degree
d of
polynomial
iff itsagreement ~90%
Theorem:
~ 90%
lines, f has
restriction
on every
line is
a univariate
with
a univariate
degree
d polynomial.
degreeresults:
d polynomial.
Weaker
[Babai, Fortnow, Lund ‘90]
[Rubinfeld Sudan
‘92]subspace]
[Line ≡ 1 dimensional
affine
[Feige, Goldwasser, Lovász, Szegedy ‘91]
≡ q points.
Stronger results:
[A. Sudan ‘96]; [Raz, Safra ‘96]
The results described in this paper indicate a possible
classification of optimization problems as to the behavior
of their approximation algorithms. Such a classification
must remain tentative, at least until the existence of
polynomial-time algorithms for finding optimal solutions
has been proved or disproved. Are there indeed O(log n)
coloring algorithms? Are there any clique finding
algorithms better than O(ne) for all e>0? Where do other
optimization problems fit into the scheme of things? What
is it that makes algorithms for different problems behave
the same way? Is there some stronger kind of reducibility
than the simple polynomial reducibility that will explain
these results, or are they due to some structural similarity
between te problems as we define them? And what other
types of behavior and ways of analyzing and measuring it
are possible?
David Johnson, 1974
NP-hard Optimization Problems
MAX-3SAT: Given 3-CNF formula , find
assignment maximizing the number of
satisfied clauses.
MAX-LIN(3): Given a linear system over
GF(2) of the form
find its largest feasible subsystem.
Approximation Algorithms
Defn: An -approximation for MAX-LIN(3) is
a polynomial-time algorithm that
computes, for each system, a feasible
subsystem of size ¸
.
( ¸ 1)
Easy Fact: 2-approximation exists.
Theorem
: If P  NP, (2-)approximation does not exists.
Common Approx. Ratios
Early History
Graham’s algorithm for multiprocessor
scheduling [approx. ratio = 2]
1971,72 NP-completeness
1974
Sahni and Gonzalez: Approximating TSP is
NP-hard
1975
FPTAS for Knapsack [IK]
1976
Christofides heuristic for metric TSP
1977
Karp’s probabilistic analysis of Euclidean TSP
1980
PTAS for Bin Packing [FL; KK]
1980-82 PTAS’s for planar graph problems [LT, B]
1966
Subsequent Developments
1988
1990
1991
1992
1992-95
MAX-SNP: MAX-3SAT is complete problem [PY]
IP=PSPACE, MIP=NEXPTIME
First results on PCPs [BFLS, FGLSS]
NP=PCP(log n,1) [AS,ALMSS]
Better algorithms for scheduling, MAXCUT [GW], MAX-3SAT,...
1995-98 Tight Lowerbounds (H97); (1+ )approximation for Euclidean TSP, Steiner Tree...
1999-now Many new algorithms and hardness
results.
2005
New simpler proof of NP=PCP(log n,1) (Dinur)
SOME NP-COMPLETE PROBLEMS
3SAT: Given a 3-CNF formula, like
decide if it has a satisfying assignment.
THEOREMS: Given
decide
if T has a proof of length · n in Axiomatic Mathematics
Philosophical meaning of P vs NP: Is there an
inherent difference between
being creative / brilliant
and
being able to appreciate creativity / brilliance?
“Feasible” computations:
those that run in polynomial
(i.e.,O(nc)) time
(central tenet of theoretical computer science)
e.g.,
time is “infeasible”
[A., Safra ‘92]
Håstad’s 3-bit PCP[A.,
Theorem
Lund, Motwani,
NP=PCP(log n, 1)
(1997) Sudan, Szegedy ’92]
nc
n
INPUT x
O(lg n) random bits
CERTIFICATE
M

Reads 3 bits;
Computes sum mod 2
O(1) bits
Accept / Reject
x is a “YES” input
 there is  s.t. Pr[ M accepts] > 1 - 
x is a “NO” input
 for every  Pr[ M rejects ] > ½ + 
(2-)-approx. to MAX-LIN(3)
) P=NP
1- _
½ +
?
INPUT x
O(lg n) random bits
M
Note: 2O(lg n) = nO(1).
) M ≡ nO(1) linear constraints
x is YES instance
) > 1- fraction satisfiable
x is NO instance
) · ½+ fraction satisfiable
Polynomial Encoding
[LFKN ‘90]
[BFL ’90]
Sequence of bits / numbers
2
4
5
7
Represent as m-variate
degree d polynomial:
2x1x2 + 4x1(1-x2) + 5x2(1-x1) + 7(1-x1)(1-x2)
Evaluate at all points in GF(q)m
Note: 2 different polynomials differ in (1-d/q)
fraction of points.
2nd Idea:
Many properties of polynomials
are locally checkable.
Program Checking [Blum Kannan ‘88]
Program Testing / Correcting [Blum, Luby, Rubinfeld ‘90]
MIP = NEXPTIME
[Babai, Fortnow, Lund ’90]
1st “PCP Theorem”
Dinur [05]’s proof uses random walks on expander graphs
instead of polynomials.
Håstad’s 3-bit Theorem (and “fourier method”)
NP = PCP(lg n, 1)
T1
c bits
1 bit
V1
T2 Raz’s Thm
S1
ck
YES instances ) 9 T1T2 Pr[Accept]
=1
vs. Pr[Accept] < 2-k/10
Verifier
Composition
V0
9 S1 S2 which V1
accepts w/ Prob ¸ 2-k/10
bits
Pr[Accept] = 1
NO instances ) 8 T1T2 Pr[Accept]
< 1-
(A few pages of
Fourier Analysis)
k bits
V2
ck
22 bits
LONG
CODING
[BGS ’95]
k
22 bits
) x is a YES instance.
Suppose
Pr[Accept] > ½ + 
S2
Sparsest Cut / Edge Expansion
G = (V, E)
S
(G) = min
| E(S, Sc)|
SµV
S
|S| < |V|/2
c- balanced separator
c(G) = min
Both NP-hard
| E(S, Sc)|
SµV
c |V| < |S| < |V|/2
|S|
|S|
Semidefinite relaxation for
c-balanced separator
|vi –vj|2/4 =1
|vi –vj|2 =0
S
+1
c(G) = min
S
| E(S, Sc)|
SµV
-1
“cut
semimetri
c”
Triangle
inequality
|S|
c |V| < |S| < |V|/2
Find unit vectors
in <n
Assign {+1, -1} to v1, v2, …, vn to minimize
(i, j) 2 E |vi –vj|2/4
Subject to i < j |vi –vj|2/4 ¸ c(1-c)n2
|vi –vj|2 + |vj –vk|2 ¸ |vi –vk|2 8 i, j, k