Skip to main content

LaTeX Fiddles

Created on February 11|Last edited on February 11

Underbraces

The layerwise gradients of a deep linear neural network have the following form:
Wklgrad w.r.t. layer k(W1,,WD)=Wk+1:later weights, adjoint   L(W)collapsed loss   W:kearlier weights\underbrace{\nabla{_{W_k}}l}_\text{grad w.r.t. layer \textit{k}}\left(W_1, \dots, W_D\right) = \underbrace{W_{k+1:}^\top}_ \text{later weights, adjoint} \ \ \ \underbrace{\nabla \mathcal{L}\left(W\right)} _\text{collapsed loss} \ \ \ \underbrace{W_{:k}}_\text{earlier weights}