Suppose we’re interested in solving an overdetermined linear least-squares problem
(1)



(2)
(3)
How small is the residual of the sketch-and-solve solution? Here’s a one bound:
Proposition 1 (Sketch-and-solve,
bound). The sketch-and-solve solution (3) satisfies the bound
Let’s prove this bound. The residual is in the range of
. Therefore, by (2),




(4)
The conclusion of Proposition 1 appears to be that the residual norm may be a factor
larger than the minimal least-squares residual
. Interestingly, this conclusion is not sharp. In fact, the residual for sketch-and-solve can only be at most
larger than optimal. This fact has been known at least since the work of Drineas, Mahoney, Muthukrishnan, and Sarlós (2007). The note of Larsen and Kolda (2022) provides a nice summary of Drineas et al.’s argument. My goal in this post is to simplify this argument even further, making it as elementary as possible.
Notation
To make our lives easier, we’ll begin by defining some notation. Let
(5)






(6)
Sketching the basis
As saw in (6), the matrix is an orthonormal basis for
. To get a sharper analysis of sketch-and-solve, we will need the following result, which shows that
is an “almost orthonormal basis”.
Lemma 2 (Sketching the basis): The matrix has nearly orthonormal columns in the sense that
(7)
(8)
This result is a pretty direct reformulation of the sketching property (2). By the variational characterization of the minimum singular value, we have



For a tall matrix , the eigenvalues of
are equal to the squared singular values of
. Applying this result to
gives

Sketch-and-Solve: Better Analysis
With Lemma 2 in hand, we are ready to state and prove our main result:
Theorem 3 (Sketch-and-solve,
bound). The sketch-and-solve solution (3) satisfies the bound
(9)
As a consequence,(10)
and(11)
Let’s prove Theorem 3. Begin by decomposing the sketch-and-solve residual :


(12)
To bound , it will help us to have a more convenient formula for
. To this end, reparametrize the sketch-and-solve least-squares problem as an optimization problem over the error
:

(13)
We now work to bound the two factors in (10). The matrix




Can We Do Better?
In the limit , Theorem 3 identifies the correct
scaling for sketch-and-solve, consistent both with numerical experiments, exact computations for Gaussian embeddings, and lower bounds for solving least-squares problems by row subselection. However, comparing results for Gaussian embeddings shows there may be room for improvement. Here is an informal statement of the Gaussian results:
Informal Theorem 4 (Gaussian sketch-and-solve): Appropriately scaled, a
Gaussian matrix is an
-sketching matrix provided
. The expected least-squares residual is
(14)
See this post of mine for a formal version of this result, its proof, and a discussion of the history of this result. We see that the bound (9) for general embeddings takes the form in the limit
, whereas the bound (14) for Gaussian embeddings scales like
. We leave it as a conjecture/open problem whether there is an improved argument for general subspace embeddings:
Conjecture 5 (Sketch-and-solve, large
): The sketch-and-solve solution (3) satisfies the bound
for an absolute constant
.
Notes: For another proof demonstrating the correction scaling for a different type of dimensionality reduction (leverage score sampling), check out this article by Raphael Meyer.