450 likes | 551 Views
Optimization in mean field random models. Johan Wästlund Linköping University Sweden. Statistical Mechanics. Each particle has a spin Energy = Hamiltonian depends on spins of interacting particles Ising model: Spins ± 1, H = # interacting pairs of opposite spin .
E N D
Optimization in mean field random models Johan Wästlund Linköping University Sweden
Statistical Mechanics • Each particle has a spin • Energy = Hamiltonian depends on spins of interacting particles • Ising model: Spins ±1, H = # interacting pairs of opposite spin
Statistical Mechanics • Spin configuration has energy H() • Gibbs measure depends on temperature T: • T→∞ random state • T→0 ground state, i.e. minimizing H()
Statistical Mechanics • Thermodynamic limit N →∞ • Average energy? (suitably normalized)
Disordered Systems • Spin glasses • AuFe random alloy • Fe atoms interact
Disordered Systems • Random interactions between Fe atoms • Sherrington-Kirkpatrick model:
Disordered Systems • Quenched random variables gi,j • S-K is a mean field model: No correlation betweeen quenched variables • NP hard to find ground state given gi,j
Computer Science • Test / evaluate heuristics for NP-hard problems • Average case analysis • Random problem instances
Combinatorial Optimization • Minimum Matching / Assignment • Minimum Spanning Tree • Traveling Salesman • Shortest Path • … • Points with given distances, minimize total length of configuration
Statistical Physics / Computer Science • Feasible solution • Cost of solution • Cost of minimal solution • Artificial parameter T • Gibbs measure • N→∞ • Spin configuration • Hamiltonian • Ground state energy • Temperature • Gibbs measure • Thermodynamic limit
Mean field models • Replica-cavity method has given good results for mean field models • Parisi solution of S-K model • The same methods can be applied to combinatorial optimization problems in mean field models
Mean field models of distance • N points • Abstract geometry • Inter-point distances given by i. i. d. random variables • Exponential distribution easiest to analyze (pseudodimension 1)
Matching • Set of edges giving a pairing of all points
Spanning tree • Network connecting all points
Traveling salesman • Tour visiting all points
Mean field limits • No normalization needed! (pseudodimension 1) • Matching: 2/12≈0.822 (Mézard & Parisi 1985, rigorous proof by Aldous 2000) • Spanning tree: (3) = 1+1/8+1/27+… ≈1.202 (Frieze 1985) • Traveling salesman: 2.0415… (Krauth-Mézard-Parisi 1989), now established rigorously!
Cavity results • Non-rigorous method • Aldous derived equivalent equations with the Poisson-Weighted Infinite Tree (PWIT)
Cavity results • Non-rigorous quantity X = cost of minimal solution – cost of minimal solution with the root removed • Define X1, X2, X3,… similarly on sub-trees • Leads to the equation • Xidistributed like X, i are times of events in rate 1 Poisson process
Cavity results • Analytically, this is equivalent to where
Cavity results • Explicit solution • Ground state energy
Cavity results • Note that the integral is equal to the area under the curve when f(u) is plotted against f(-u) • In this case, f satisfies the equation
K-L matching • Similarly, the K-L matching problem leads to the equations: • has rate K and has rate L • min[K] stands for K:th smallest
K-L matching • Shown by Parisi (2006) that this system has an essentially unique solution • The ground state energy is given by where x and y satisfy an explicit equation • For K=L=2, this equation is • Unfortunately the cavity method is not rigorous
The exponential bipartite assignment problem • Exact formula conjectured by Parisi (1998) • Suggests proof by induction • Researchers in discrete math, combinatorics and graph theory became interested • Generalizations…
Generalizations • by Coppersmith & Sorkin to incomplete matchings • Remarkable paper by M. Buck, C. Chan & D. Robbins (2000) • Introduces weighted vertices • Extremely close to proving Parisi’s conjecture!
Weighted assignment problems • Weights 1,…,m, 1,…, n on vertices • Edge cost exponential of rate ij • Conjectured formula for the expected cost of minimum assignment • Formula for the probability that a vertex participates in solution (trivial for less general setting!)
a3 a1 a2 The Buck-Chan-Robbins urn process • Balls are drawn with probabilities proportional to weight
Proofs of the conjectures • Two independent proofs of the Parisi and Coppersmith-Sorkin conjectures were announced on March 17, 2003 (Nair, Prabhakar, Sharma and Linusson, Wästlund)
Annealing • Powerful idea: Let T→0, forcing the system to converge to its ground state • Replica-cavity approach • Simulated annealing meta-algorithm (optimization by random local moves)
In the mean field model: Underlying rate 1 variables Yi • ri plays the same role as T • Local temperature • Associate weight to vertices rather than edges
Cavity/annealing method • Relax by introducing an extra vertex • Let the weight of the extra vertex go to zero • Example: Assignment problem with 1=…=m=1, 1=…=n=1, and m+1 = • p = P(extra vertex participates) • p/n = P(edge (m+1,n) participates)
Annealing • p/n = P(edge (m+1,n) participates) • When →0, this is • Hence • By Buck-Chan-Robbins urn theorem,
Annealing • Hence • Inductively this establishes the Coppersmith-Sorkin formula
Results with annealing • Much simpler proofs of Parisi, Coppersmith-Sorkin, Buck-Chan-Robbins formulas • Exact results for higher moments • Exact results and limits for optimization problems on the complete graph
The 2-dimensional urn process • 2-dimensional time until k balls have been drawn
Limit shape as n→∞ • Matching: • TSP/2-factor:
Mean field TSP • If the edge costs are i.i.d and satisfy P(l<t)/t→1 as t→0 (pseudodimension 1), then as n →∞, • A. Frieze proved that whp a 2-factor can be patched to a tour at small cost
Future work • Explain why the cavity method gives the same equation as the limit shape in the urn process • Establish more detailed cavity predictions • Use proof method of Nair-Prabhakar-Sharma in more general settings