Nearly two weeks ago Indian-American mathematician
Vinay Deolalikar, the employee of HP Labs, sent a letter to a group of mathematicians asking them to proof-read his attempt to prove
P ≠ NP. The pre-print was eventually distributed over the Internet, so a lot of people became aware. Later two major flaws were found in the proof, so it is now considered wrong. Nevertheless, Vinay has removed the paper from his page and commented that he fixed all the issues and is going to submit a paper to the journal review. You can download the original paper from my web-site [
Deolalikar, 2010].
We'll see if he is bluffing or not. In any case, it is very pleasant to me that the main role in the proof is played by... the graphical probabilistic model framework. Yeah, those graphical models we use in computer vision! It was surprising indeed, since the problems in computational complexity are usually discrete and well-posed. So, this post is devoted to understanding why a graphical model emerges in the proof.
First, I am going to review some key points of the proof. It is far from being exhaustive, just some rough sketches that are critical for understanding. Then I report on the flaws found in the proof, and then I end up with a methodological note on the whole P vs. NP problem.
k-SAT and d1RSB Phase
In order to prove P ≠ NP, it is enough to show the absence of a polynomial algorithm for some problem from
the class NP, e.g. for some
NP-complete problem (which are the "hardest" problems in NP).
The boolean satisfiability problem (SAT) addresses the issue of checking if there is at least one assignment of the boolean variables in the formula, represented in conjunctive normal form (CNF), which makes it TRUE. The k-satisfiability problem (k-SAT) is a particular case of SAT, where all the clauses in the CNF have order
k. For example, (
x˅
y)&(¬
x˅
y)&(
x˅¬
y)&(¬
x˅¬
y) is an instance of 2-SAT which has the answer 'NO', since any boolean values of (
x, y) makes the formula false. (
x˅
y˅
z)&(¬
x˅
u˅
v)&(¬
x˅
u˅¬
z) is a 3-SAT over 5 variables (x, y, z, u, v), which is satisfiable, e.g. on the input (1,0,0,1,0). 2-SAT is in P, but k-SAT is NP-complete whenever
k > 2. The Deolalikar's idea is to show that k-SAT (with
k > 8) is outside of P, which (if true) proves that there is a separation between P and NP.
The space of possible k-SAT solutions is analysed. Let
m be the number of clauses in the CNF,
n be the number of variables. Consider the ratio
α =
m/
n. In the border case
m = 1 (
α is small), k-SAT is always true, there are usually a lot of inputs that satisfy the CNF. Consider an ensemble of random k-SATs (the following is statistical reasoning). With the growth of
α, the probability of the CNF to be satisfiable decreases. When a certain threshold
αd is reached, the "true" solutions set breaks into clusters. This stage is known in statistical mechanics as
dynamical one step replica symmetry breaking (d1RSB) phase. Moving further, we make the problem completely unsatisfiable. It turns out that d1RSB stage subproblems are the most challenging of the all k-SAT problems. The effect could be observed only if
k > 8, that's why such problems are used in the proof. [
Deolalikar, 2010, Chapter 6]
FO(PFP) and FO(LFP)
In the
finite model theory there are recurrent extensions of the first-order logic. The predicate P
i(
x) is evaluated as some second-order function
φ(P
i-1, x), where x is a
k-element vector, P
0(
x) ≡ false. In the general case, either there is a fixed point, or is P
i looping. For example, if
φ(P
, x) ≡ ¬P(
x), then P
0(
x) ≡ false, P
1(
x) ≡ true, P
2(
x) ≡ false, etc. Here
x is meaningless, but it is not always the case. Consider the following definition:
φ(P
, x) ≡ max(
x) ˅ P(
x) // recall
x is a vector, in the boolean case 'max' is the same as 'or'. If
x contains at least one non-zero element, P
0(
x) = false, P
1(
x) = true, P
2(
x) = true, etc. Otherwise, P
i(
0) = false for all
i. In the case of looping, let's redefine the fixed point to be constantly false.
FO(PFP) is a class of problems of checking if there will be a loop for some
x, or a fixed point. They are actually very difficult problems. FO(PFP) is equal to the whole
PSPACE. Suppose now that
φ is monotonic on P. It means that P(
x) only appears in the formula with even number of negations before it (or zero, as in the second example). This means that once P
i(
x) is true, P
j(
x) will be true for any
j >
i. This reduces the class to
FO(LFP), which is proven to be equal to the class
P. So, the global problem now is to show that k-SAT is outside of the class FO(LFP).
ENSP: the graphical model for proving P ≠ NP
So how a graphical model emerges here? Graphical model is a way to describe a multivariate probability distribution, i.e. dependencies of covariates in the distribution. Recall now
the definition of NP. If we somehow guess the certificate, we are done (i.e. have a polynomial algorithm). If the space of certificates (possible solutions, in terms of Deolalikar) is simple, we can probably get a polynomial algorithm. What is simple? This means that the distribution is parametrizable with a small amount of parameters (c.f.
intrinsic dimensionality), which allows us to traverse the solution space efficiently. This is twofold. First, the distribution is simple if it has a limited support, i.e. all the probability measure is distributed among a limited number of points of the probabilistic space. Second, it is simple if the covariates are as much conditionally independent as possible. In the ideal case, if all the groups of covariates are independent (recall that
pairwise and mutual independence do not subsume each other!), we need only
n parameters to describe the distribution (
n is the number of covariates), while in general case this number is exponential. See [
Deolalikar, 2010, p. 6] for examples.
How to measure the degree of conditional independence? Yes, to build a graphical model. It is factorized to cliques according to
Hammersley-Clifford theorem. Larger cliques you get, stronger dependency is. When the largest cliques are k-interactions, the distribution can be parametrized with
n2
k independent parameters. Finally, Vinay shows that FO(LFP) can deal only with the distributions parametrizable with 2
poly(log n) parameters, which is not the case for d1RSB k-SAT (its solution space is too complicated).
In order to show it strictly, Deolalikar introduces a tailored graphical model describing LO(LPF) iterative process, the Element-Neighborhood-Stage Product model (ENSP):
There are two types of sites: corresponding to the variables on the stages of LFP (small circles) and corresponding to the elements of witty neighbourhood system (some closure over
Gaifman graph; big blue circles). When a variable is assigned 0 or 1, it is painted with red or green respectively. The last column corresponds to the fixed point, all the variables are assigned. Thus, this model is easily factorizable to the small cliques, and it cannot describe the imaginable LFP process for some k-SAT. See [
Deolalikar, 2010, Chapter 8] for the details.
So what's the problem?
Neil Immerman, an expert in the finite model theory, noticed two major flaws. The first one is that Vinay actually model only monadic LFP, which is not equal to P, as assumed. He thus proved that there is a gap between NP and some subclass of P, which is not obligatory equals P. The second major flow deals with modelling k-SAT as a logical structure. I do not fully understand this step, but some order-invariance assumptions are wrong. Here is a discussion in
Lipton's blog. According to it, the flaws are really fatal.
The third way
It seems that it is really difficult to prove P ≠ NP. The most of scientists think that P = NP is improbable, and that's why the hypothesis P ≠ NP is generally adopted now. But there is one more option: neither P = NP nor P ≠ NP is provable. As every mathematical theory, computational complexity is a set of axioms and statements, which are usually could be proved to be correct or not. But there are sometimes some statements formulated in terms of the theory, which could not. Moreover, according to
the first Gödel's incompleteness theorem, in any mathematical theory based on our natural understanding of natural numbers, there is either a statement that could be proved both true or false, or an unprovable one. I know no actual reasons why this could not be the case for P ≠ NP (although I have feelings there are some since it is not usually taken into account).
Suppose it is really so. This would mean that the whole theory should be reconsidered. Some axioms or definitions will change to make this fact provable. But may be anything else will become unprovable. Who knows...