SU-CS229 Distribution Sheet
Last edited: October 10, 2025Here’s a bunch of exponential family distributions. Recall:
\begin{equation} p\qty(x;\eta) = b\qty(x) \exp \qty(\eta^{T}T\qty(x) - a\qty(\eta)) \end{equation}
normal, berunouli, posisson, binomial, negative binomial, geometric, chi-squared, exponential are all in
normal distribution
\(\mu\) the mean, \(\sigma\) the variance
\begin{equation} p\qty(x;\mu, \Sigma) = \frac{1}{\qty(2\pi)^{\frac{|x|}{2}} \text{det}\qty(\Sigma)^{\frac{1}{2}}} \exp \qty(-\frac{1}{2} \qty(x-\mu)^{T}\Sigma^{-1}\qty(x-\mu)) \end{equation}
\begin{equation} p\qty(x; \mu, \sigma) = \frac{1}{\sqrt{2\pi\sigma^{2}}} \exp \qty({ \frac{-(x-u)^{2}}{2 \sigma^{2}}}) \end{equation}
\begin{equation} \mathbb{E}[x] = \mu \end{equation}
\begin{equation} \text{Var}\qty [x] = \sigma^{2} \end{equation}
This is exponential family distribution. For \(\sigma^{2} = 1\):
SU-CS229 OCT292025
Last edited: October 10, 2025xavier initialization
Last edited: October 10, 2025An neural network initialization scheme that tries to avoid Vanishing Gradients.
Consider \(Wx\) step in a neural network:
\begin{equation} o_{i} = \sum_{j=1}^{n_{\text{in}}} w_{ij} x_{j} \end{equation}
The variance of this:
\begin{equation} \text{Var}\qty [o_{i}] = n_{\text{in}} \sigma^{2} v^{2} \end{equation}
expectation maximization
Last edited: October 10, 2025Sorta like “distribution-based k-means clustering”. Jensen’s Inequality guarantees convergence (i.e. each parameter will converge to the maximum possible parameter).
constituents
requirements
Two steps:
e-step
“guess the value of \(z^{(i)}\); soft guesses of cluster assignments”
\begin{align} w_{j}^{(i)} &= p\qty(z^{(i)} = j | x^{(i)} ; \phi, \mu, \Sigma) \\ &= \frac{p\qty(x^{(i)} | z^{(i)}=j) p\qty(z^{(i)}=j))}{\sum_{l=1}^{k}p\qty(x^{(i)} | z^{(i)}=l) p\qty(z^{(i)}=l))} \end{align}
Where we have:
- \(p\qty(x^{(i)} |z^{(i)}=j)\) from the Gaussian distribution, where we have \(\Sigma_{j}\) and \(\mu_{j}\) for the parameters of our Gaussian \(j\).
- \(p\qty(z^{(i)} =j)\) is just \(\phi_{j}\) which we are learning
These weights \(w_{j}\) are how much the model believes it belongs to each cluster.
Kosaraju's Algorithm
Last edited: October 10, 2025Finding strongly connected components.
- Do DFS starting at random node to create a DFS forest (i.e. disconnected DFS trees), keeping track of finishing times
- Reverse all edges in the graph
- Do DFS again to create another DFS forest, start DFS procedure with the vertices with the largest finishing time in step 1
- strongly connected components are the trees that changed from the first in the second DFS graph
main idea
Recall that finishing time helps us track topological sort. After reversing edges, the condensation graph node with the largest finish time has no edges going out (i.e. it was originally the first in the topological sort); if we run DFS there, we will find exactly that component. After DFSing there, we will have “removed” that component (having visited it), and repeat.
