# PCP theorem

Not to be confused with Post correspondence problem.

In computational complexity theory, the PCP theorem (also known as the PCP characterization theorem) states that every decision problem in the NP complexity class has probabilistically checkable proofs (proofs that can be checked by a randomized algorithm) of constant query complexity and logarithmic randomness complexity (uses a logarithmic number of random bits).

The PCP theorem says that for some universal constant K, for every n, any mathematical proof of length n can be rewritten as a different proof of length poly(n) that is formally verifiable with 99% accuracy by a randomized algorithm that inspects only K letters of that proof.

The PCP theorem is the cornerstone of the theory of computational hardness of approximation, which investigates the inherent difficulty in designing efficient approximation algorithms for various optimization problems. It has been described by Ingo Wegener as "the most important result in complexity theory since Cook's Theorem"[1] and by Oded Goldreich as "a culmination of a sequence of impressive works [...] rich in innovative ideas".[2]

## Formal statement

The PCP theorem states that

NP = PCP[O(log n), O(1)].

## PCP and hardness of approximation

An alternative formulation of the PCP theorem states that the maximum fraction of satisfiable constraints of a constraint satisfaction problem is NP-hard to approximate within some constant factor.

Formally, for some constants K and α < 1, the following promise problem (Lyes, Lno) is an NP-hard decision problem:

• Lyes = {Φ: all constraints in Φ are simultaneously satisfiable}
• Lno = {Φ: every assignment satisfies fewer than an α fraction of Φ's constraints},

where Φ is a constraint satisfaction problem over Boolean alphabet with at most K variables per constraint.

As a consequence of this theorem, it can be shown that the solutions to many natural optimization problems including maximum boolean formula satisfiability, maximum independent set in graphs, and the shortest vector problem for lattices cannot be approximated efficiently unless P = NP. These results are sometimes also called PCP theorems because they can be viewed as probabilistically checkable proofs for NP with some additional structure.

## History

The PCP theorem is the culmination of a long line of work on interactive proofs and probabilistically checkable proofs. The first theorem relating standard proofs and probabilistically checkable proofs is the statement that NEXPPCP[poly(n), poly(n)], proved by Babai, Fortnow & Lund (1990).

### Etymology of the name "PCP theorem"

The notation PCPc(n), s(n)[r(n), q(n)] is explained at Probabilistically checkable proof. The notation is that of a function that returns a certain complexity class. See the explanation mentioned above.

The name of this theorem (the "PCP theorem") probably comes either from "PCP" meaning "probabilistically checkable proof", or from the notation mentioned above – (or both).

### History after the first theorem [in 1990]

Subsequently, the methods used in this work were extended by Babai, Lance Fortnow, Levin, and Szegedy in 1991 (Babai et al. 1991), Feige, Goldwasser, Lund, Safra, and Szegedy (1991), and Arora and Safra in 1992 (Arora & Safra 1998) to yield a proof of the PCP theorem by Arora, Lund, Motwani, Sudan, and Szegedy in 1992 (Arora et al. 1998).

The 2001 Gödel Prize was awarded to Sanjeev Arora, Uriel Feige, Shafi Goldwasser, Carsten Lund, László Lovász, Rajeev Motwani, Shmuel Safra, Madhu Sudan, and Mario Szegedy for work on the PCP theorem and its connection to hardness of approximation.

In 2005 Irit Dinur discovered a different proof of the PCP theorem, using expander graphs.[3]

### Quantum analog of the PCP Theorem

In 2012, Thomas Vidick and Tsuyoshi Ito published a result[4] that showed a "strong limitation on the ability of entangled provers to collude in a multiplayer game." This could be a step toward proving the quantum analogue of the PCP theorem, since when the result[4] was reported in the media,[5][6] professor Dorit Aharonov called it "the quantum analogue of an earlier paper on multiprover interactive proofs" that "basically led to the PCP theorem".[6]

## Notes

1. ^ Ingo Wegener (2005). Complexity Theory: Exploring the Limits of Efficient Algorithms. Springer. p. 161. ISBN 978-3-540-21045-0.
2. ^ Oded Goldreich (2008). Computational Complexity: A Conceptual Perspective. Cambridge University Press. p. 405. ISBN 978-0-521-88473-0.
3. ^ See the 2005 preprint, ECCC TR05-046. The authoritative version of the paper is Dinur (2007).
4. ^ a b Ito, Tsuyoshi; Vidick, Thomas. "A multi-prover interactive proof for NEXP sound against entangled provers".
5. ^ Hardesty, Larry (2012-07-30). "MIT News Release: 10-year-old problem in theoretical computer science falls". MIT News Office. Archived from the original on 2012-08-10. Retrieved 2012-08-10. Interactive proofs are the basis of cryptographic systems now in wide use, but for computer scientists, they're just as important for the insight they provide into the complexity of computational problems.
6. ^ a b Hardesty, Larry (2012-07-31). "10-year-old problem in theoretical computer science falls". MIT News Office. Archived from the original on 2012-08-10. Retrieved 2012-08-10. Dorit Aharonov, a professor of computer science and engineering at Hebrew University in Jerusalem, says that Vidick and Ito’s paper is the quantum analogue of an earlier paper on multiprover interactive proofs that “basically led to the PCP theorem, and the PCP theorem is no doubt the most important result of complexity in the past 20 years.” Similarly, she says, the new paper “could be an important step toward proving the quantum analogue of the PCP theorem, which is a major open question in quantum complexity theory.”