410 likes | 585 Vues
New Insights into Semidefinite Programming for Discrete Optimization. Moses Charikar Princeton University. Optimization Problems. Shortest paths Minimum cost network Scheduling, Load balancing Graph partitioning problems Constraint satisfaction problems. Approximation Algorithms.
E N D
New Insights into Semidefinite Programming for Discrete Optimization Moses Charikar Princeton University
Optimization Problems • Shortest paths • Minimum cost network • Scheduling, Load balancing • Graph partitioning problems • Constraint satisfaction problems
Approximation Algorithms • Many optimization problems NP-hard • Alternate approach: heuristics with provable guarantees • Guarantee: Alg(I) OPT(I) (maximization)Alg(I) OPT(I) (minimization) • Complexity theory gives bounds on best approximation ratios possible
i ¢ x m n c b A ¸ x Mathematical Programming approaches • Sophisticated tools from convex optimization • e.g. Linear programming • Can find optimum solution in polynomial time
Relax and Round • Express solution in terms of decision variables, typically {0,1} or {-1,+1} • Feasibility constraints on decision variables • Objective function • Relax variables to get mathematical program • Solve program optimally • Round fractional solution
LP is a widely used tool in designing approximation algorithms • Interpret variables values as probabilities, distances, etc. integer solutions fractional solutions
Distance function from LPs • [Leighton, Rao ’88] Distance function d. • Triangle ineq.:d(a, b) + d(b, c) ¸ d(a, c) d = 1 d = 0 d = 0 1 a c 0 0 b
Quadratic programming • Linear expressions in xi xj ? • NP-hard • Workaround: Mij = xi xj • What can we say about M ? • M is positive semidefinite (psd) • Can add psd constraint • Semidefinite programming • Can solve to any desired accuracy
Positive Semidefinite Matrices • M is psd iff • xT M x 0 for all x • All eigenvalues of M are non-negative • M = VT V (Cholesky decomposition) • Mij = vi vj
Vector Programming • Variables are vectors • Linear constraints on dot products • Linear objective on dot products
Max-Cut • Given graph G • Partition vertices into two sets • Maximize number of edges cut • Random solution cuts half the edges • Nothing better known until Goemans-Williamson came along !
2 2 ( ( ) ) ¡ ¡ x v v x X X i i j j m m a a x x 4 4 ( ( ) ) E E i i j j 2 2 ; ; f l l 2 i 1 f l l i 1 ¢ v v o r a = i i x o r a = i Relaxation for Max Cut
SDP solution • Geometric embedding of vertices • Hyperplane rounding
[ ] = µ E t c u ¼ i ¸ m n ( ( ) ) = µ S D P 1 2 ¡ µ c o s Rounding SDP solution • Pick random vector r • Partition vertices according to sign(vi·r) • Prob[(i,j) cut] = ij / • Contribution of (i,j) to SDP = (1-cos ij)/2 • 0.878 approximation
Can we do better ? • Better analysis ? rounding algorithm ? • [Karloff ‘97] guarantee for random hyperplane cannot be improved. • [Feige, Schechtman ’01]SDP value can differ from optimal by 0.878
An Improved Bound ? • Add constraints to the relaxation. • -inequality constraints: • (vi –vj)2 + (vj –vk)2 (vi – vk)2 • [Feige, Schechtman ’01]showed gap for SDP with -inequalities, slightly better than 0.878
SDP applications • DiCut, Max k-Cut • Constraint satisfaction problems2-SAT, 3-SAT • Graph coloring • …
j i Sparsest Cut uniform demands ( ) ± S T ; i m n j j j j S T ¢ S T
( ) d f h d i i j t t c c - c a p a c y o e e g e i i i j j j ; X X ( ) d d d f h d i i j t - e m a n o e p a r c i j i i j j ; ( S ) T i j E i j 2 2 2 ; ; S T i j 2 2 ; ( ) ± S T i ; m n i m n j j j j S T ¢ Sparsest Cut S T
i j ( ) ± i j ; Cut Metric 1 0 0 S T Use relaxations of cut metrics
Approximate cut metrics • How well can relaxed metrics be mapped into cut metrics ? • Metrics from LPs: log n distortion gives log n approximation[Bourgain ’85] [LLR ’95] [AR ’95] • SDP with -inequalities ? • geometry of l22 metrics • Goemans-Linial conjecture:l22 metrics embed into l1 with constant distortion.
p l o g n Arora-Rao-Vazirani • [ARV ’04] • Breakthrough for SDPs with -inequalities • approximation for balanced cut and sparsest cut
ARV-Separation Theorem • [Arora, Rao, Vazirani ’04, • Lee ‘05]: • Unit vectors visatisfy triangle inequalities: • |vi – vj|2 + |vj – vk|2¸ |vi – vk|2 • (and a spreading constraint) • ) sets S and T, that: • =(1/(log n)½) separated • contain a const. fraction of all vertices (each) S T
p p l l o o g g n n Applications • min unCut • approximation [ACMM ’05] • Min 2-CNF deletion • approximation [ACMM ’05] • Directed analog of ARV separation lemma
Directed Distances • Choose an arbitrary unit vector v0 • Define directed symmetric semimetric d as follows: d(vi , vj) = |vi – vj|2 + 2hvi – vj , v0i = 2hv0 + vi , v0 – vji = |vi – vj|2 + |v0 – vj|2 – |v0 – vi|2 d = 1 d = 0 d = 0 d = 0 true false vi = v vi = -v
Applications • Arrangement problems • Minimum Linear Arrangement [CHKR ’06] [FL ’06] • Embedding in d-dimensions[CMM ’07] • Graph coloring • O(n0.2)coloring of 3-colorable graphs[ACC ’06]
8 ( ) d 2 1 3 3 1 7 + ´ x x m o 1 4 > > > ( ) d 1 6 4 1 7 + < ´ x x m o 3 2 > : : : > > ( ) d : 5 3 9 1 7 + ´ x x m o 1 9 Unique Games • Linear equations mod p • 2 variables per equation • maximize number of satisfied constraints • In every constraint, for every value of one variable, unique value of other variable satisfies the constraint. • If 99% of equations are satisfiable, can we satisfy 1% of them ?
Unique Games Conjecture • [Khot ’02]Given a Unique Games instance where 1-fraction of constraints is satisfiable, it is NP-hard to satisfy evenfraction of all constraints. (for every constant positive and and sufficiently large domain size k).
Implications of UGC • 2 is best possible for Vertex cover [KR ’03] • 0.878 is best possible for Max Cut [KKMO ’04] [MOO ’05] • (1) for sparsest cut(1) for min 2CNF deletion[CKKRS ’05] [KV ’05]
Algorithms for Unique Games • Domain size k, OPT = 1- • Random solution satisfies 1/k • Non-trivial results only for = 1/poly(k)[AEH ’01] [Khot ’02] [Trevisan ’05] [GT ’06] 1 - 0 1
³ ´ p l k O 1 ¡ " o g " ¡ ¡ ¢ k ¡ 2 " New results for Unique Games • [CMM ’05] • Given an instance where 1-fraction of constraints is satisfiable, we satisfy • We can also satisfy:
³ ´ p l k O 1 ¡ " o g = l k 1 1 0 c o g = 2 1 5 ( ) k O 1 = ( ) ¡ 2 1 ¡ ¡ ¡ " " " k k New results for Unique Games • Algorithms cover the entire range of .
Seems distant from UGC setting • Optimal if UGC is true ![KKMO ’05] [MOO ’05] • Any improvement will disprove UGC 1 - 0 1
p h l k G i t t > ( ¢ v e n a g u o g " ¡ k ¡ 2 " p [ ] h l k ? P i t > ¢ w a s r g v o g p ( ) l k O 1 ¡ " o g Matching upper and lower bounds ? g Gaussian random vector v u u · v = 1
If pigs could whistle … • UGC seems to predict limitations of SDPs correctly • UGC based hardness for many problems matching best SDP based approximation • UGC inspired constructions of gap examples for SDPs • Disproof of Goemans-Linial conjecturel22 metrics do not embed into l1 with constant distortion. [KV ’05]
Is UGC true ? • Points to limitations of current techniques • Focuses attention on common hard core of several important optimization problems • Motivates development of new techniques
Approaches to disproving UGC • Focus on possibly easier problems • Max Cut: • OPT = 1-, beat 1-1/2[GW ‘94] • Max k-CSP: • constraints are ANDs of k literals • maximize #satisfied constraints • Beat k/2k [ST ‘06] [CMM ‘07] • Distinguish between 1/k and 1/2k satisfiable
Approaches to disproving UGC • Lifting procedures for SDPs • Lovasz-Schrijver, Sherali-Adams, Lasserre • Simulate products of k variables • Can we use them ?
Moment matrices • SDP solution gives covariance matrix M • There exist normal random variables with covariances Mij • Basis for SDP rounding algorithms • There exist {+1,-1} random variables with covariances Mij/log n • Is something similar possible for higher order moment matrices ?
Concluding thoughts • Fascinating questions • Algorithms require geometric insights • Is the geometry intrinsic to these problems ? • Many mysterious connections and unsolved problems