Convex Optimization Index
Last edited: February 2, 2026EE364A.stanford.edu
Lecture
Maximum Likelihood Estimation with Convex Optimization
Last edited: February 2, 2026motivation
Consider Generic Maximum Likelihood Estimate.
- parametric distribution estimation: suppose you have a family of densities \(p_{x}\qty(y)\), with parameter \(x\)
- we take \(p_{x}\qty(y) = 0\) for invalid values of \(x\)
maximum likelihood estimation: choose \(x\) to maximize \(p_{x}\qty(y)\) given some dataset \(y\).
linear measurement with IID noise
Suppose you have some kind of linear noise model:
\begin{equation} y_{i} = a_{i}^{T}x + v_{i} \end{equation}
where \(v_{i}\) is IID noise, and \(a^{T}_{i}\) is the model. We can write \(y\) probabilistically as:
norm approximation
Last edited: February 2, 2026Consider an error minimization task \(\min \norm{Ax - b}\) (\(Ax\) as the “predictions”, and \(b\) is the “data”). Some interpretation—
- approximation: \(Ax^{*}\) is the best approximation of the vector \(b\) by linear combinations of columns of \(A\)
- geometric: \(Ax^{*}\) is a point in \(\mathcal{R}\qty(A)\) closest to \(b\)
- estimation: linear measurement model \(y = Ax + v\)
- you took a measurement \(y\), \(A\) is the theoretical measurement, \(v\) is the measurement error
- implausibility of making \(v\) error is \(\norm{v}\)
- given \(y = b\) (what you measured), most plausible \(x\) is \(x^{*}\)
- optimal design: \(x\) are design variables, \(Ax\) is the result; \(x^{*}\) is the design that best approximates desired \(b\)
Penalty Function Approximation
Suppose you are optimizing some design \(x\) with respect to some dynamics \(A\). Suppose your design residual is \(r = Ax - b\). And you have some kind of penalty function to describe how comfortable you are with various errors: \(\phi\qty(r_1) + … + \phi\qty(r_{n})\).
perturbation analysis
Last edited: February 2, 2026Let the unpreturb problem be:
\begin{align} \min_{x}\quad & f_{0}\qty(x) \\ \textrm{s.t.} \quad & f_{i}\qty(x) \leq 0, i = 1 \dots m \\ & h_{i}\qty(x) = 0, i = 1 \dots p \end{align}
Preturbed on is just:
\begin{align} \min_{x}\quad & f_{0}\qty(x) \\ \textrm{s.t.} \quad & f_{i}\qty(x) \leq u_{i} \\ & h_{i}\qty(x) = v_{i} \end{align}
Global Tightness
So we can get a lower bound:
\begin{equation} p^{*}\qty(u,v) \geq g\qty(\lambda^{*}, v^{*}) - u^{T} \lambda^{*} - v^{T}\lambda^{*} \end{equation}
Stochastic Robust Least Squares
Last edited: February 2, 2026Consider the Robust Approximation problem:
Robust Approximation
Minimize \(\norm{Ax - b}\) with uncertain \(A\); two approaches—
- stochastic: minimize \(\mathbb{E}\norm{Ax -b }\)
- worst-case: set \(\mathcal{A}\) of possible values of \(A\), minimize \(\text{sup}_{A \in A} \norm{Ax - b}\)
