In compressed sensing, the nullspace property gives necessary and sufficient conditions on the reconstruction of sparse signals using the techniques of ℓ 1 -relaxation. The term "nullspace property" originates from Cohen, Dahmen, and DeVore. The nullspace property is often difficult to check in practice, and the restricted isometry property is a more modern condition in the field of compressed sensing.
The non-convex ℓ 0 -minimization problem,
min x ∥ x ∥ 0 subject to A x = b ,
is a standard problem in compressed sensing. However, ℓ 0 -minimization is known to be NP-hard in general. As such, the technique of ℓ 1 -relaxation is sometimes employed to circumvent the difficulties of signal reconstruction using the ℓ 0 -norm. In ℓ 1 -relaxation, the ℓ 1 problem,
min x ∥ x ∥ 1 subject to A x = b ,
is solved in place of the ℓ 0 problem. Note that this relaxation is convex and hence amenable to the standard techniques of linear programming - a computationally desirable feature. Naturally we wish to know when ℓ 1 -relaxation will give the same answer as the ℓ 0 problem. The nullspace property is one way to guarantee agreement.
An m × n complex matrix A has the nullspace property of order s if for all index sets S with s = | S | ≤ n we have that: ∥ η S ∥ 1 < ∥ η S C ∥ 1 for all η ∈ ker A ∖ { 0 } .
The following theorem gives necessary and sufficient condition on the recoverability of a given s -sparse vector in C n . The proof of the theorem is a standard one, and the proof supplied here is summarized from Holger Rauhut.
Theorem: Let A be a m × n complex matrix. Then every s -sparse signal x ∈ C n is the unique solution to the ℓ 1 -relaxation problem with b = A x if and only if A satisfies the nullspace property with order s .
Proof: For the forwards direction notice that η S and − η S C are distinct vectors with A ( − η S C ) = A ( η S ) by the linearity of A , and hence by uniqueness we must have ∥ η S ∥ 1 < ∥ η S C ∥ 1 as desired. For the backwards direction, let x be s -sparse and z another (not necessary s -sparse) vector such that z ≠ x and A z = A x . Define the (non-zero) vector η = x − z and notice that it lies in the nullspace of A . Call S the support of x , and then the result follows from an elementary application of the triangle inequality: ∥ x ∥ 1 ≤ ∥ x − z S ∥ 1 + ∥ z S ∥ 1 = ∥ η S ∥ 1 + ∥ z S ∥ 1 < ∥ η S C ∥ 1 + ∥ z S ∥ 1 = ∥ − z S C ∥ 1 + ∥ z S ∥ 1 = ∥ z ∥ 1 , establishing the minimality of x . ◻