Houjun Liu

SU-CS205L FEB042025

“regularization”: removing columns (i.e. explicitly making things roughly degenerate to fit more generally)

“tall full-rank matrix” (i.e. let \(A\) be \(m \times n\), such that \(m \geq n\)).

Solving such full-rank linear systems—

\begin{equation} Ac = b \implies U \Sigma V^{T} c = b \implies \Sigma \qty(V^{T}c) = U^{T} b \end{equation}

We can now solve this above by simply dividing by the nonzero singular values of each row (since we are full rank, this is true). The last equations are zero.


“classic” method for solving least squares—

\begin{equation} A^{T}A c = A^{T} b \end{equation}

this only works if we know \(Ac = b\) (that the problem is solvable.)


residuals based method for solving the problem—

\(b \in \text{range}\ A\), then \(Ac = b\). However, generally, only \(r = b- Ac\) is always true. So then the goal is to minimize residual:

\begin{equation} \min_{c} \qty(b - Ac) \end{equation}

Notice this is a vector! So to actually minimize it, we need to pick a norm. By convention we use 2 norms:

\begin{equation} \min_{c} \Vert \qty(b - Ac) \Vert_{2} \end{equation}


weighted least squares: some equations maybe more important than others—scaling the entries in the residuals (prior taking the norm); so we are actually trying to minimize \(\Vert Dr \Vert_{2}\) with diagonal \(D\), in particular meaning we have:

\begin{equation} Dr = Db - DAc \end{equation}

Suppose you are trying to minimize \(\Vert r \Vert^{2}_{2}\) to get rid of the square root (and so we can just foil and get a matrix equation):

\begin{equation} c^{T} A^{T} A c - 2b^{T} Ac \end{equation}

for weighting:

\begin{equation} c^{T}A^{T} D^{2}Ac - 2b^{T} D^{2} Ac \end{equation}

Critical point: \(\nabla f = 0\) ; any \(c\) that simultaneously solves all equations is a critical point.

The Hessian is the Jacobian of the gradient; note that order matters.

by then finding the critical point on \(c\), and setting \(D = I\), we obtain the normal normal equations from above.


For multiplying by diagonal matricies, systems with unique solutions


Since the eigenvalues of \(A^{T}A\) is double than that of \(A\) itself, the condition numbers:

\begin{equation} \frac{\sigma^{2}_{\text{max}}}{\sigma^{2}_{\min}} \end{equation}

its takes squared the amount of precision to do what you want to do.

For solving least squares, we use a QR factorization, where QR is R upper triangualar and Q^T Q = I.

intutition: Qs—orthogonal columns from \(A\); Rs—any part that we chopped away from the orthogonmalization.

\(Q^{T}Q = I\); this is NOT true in the other direction.