Welcome to Randal Douc's wiki

A collaborative site on maths but not only!

User Tools

Site Tools


$$ \newcommand{\arginf}{\mathrm{arginf}} \newcommand{\argmin}{\mathrm{argmin}} \newcommand{\argmax}{\mathrm{argmax}} \newcommand{\asconv}[1]{\stackrel{#1-a.s.}{\rightarrow}} \newcommand{\Aset}{\mathsf{A}} \newcommand{\b}[1]{{\mathbf{#1}}} \newcommand{\ball}[1]{\mathsf{B}(#1)} \newcommand{\bproof}{\textbf{Proof :}\quad} \newcommand{\bmuf}[2]{b_{#1,#2}} \newcommand{\card}{\mathrm{card}} \newcommand{\chunk}[3]{{#1}_{#2:#3}} \newcommand{\convprob}[1]{\stackrel{#1-\text{prob}}{\rightarrow}} \newcommand{\Cov}{\mathbb{C}\mathrm{ov}} \newcommand{\CPE}[2]{\PE\lr{#1| #2}} \renewcommand{\det}{\mathrm{det}} \newcommand{\dimlabel}{\mathsf{m}} \newcommand{\dimU}{\mathsf{q}} \newcommand{\dimX}{\mathsf{d}} \newcommand{\dimY}{\mathsf{p}} \newcommand{\dlim}{\Rightarrow} \newcommand{\e}[1]{{\left\lfloor #1 \right\rfloor}} \newcommand{\eproof}{\quad \Box} \newcommand{\eremark}{</WRAP>} \newcommand{\eqdef}{:=} \newcommand{\eqlaw}{\stackrel{\mathcal{L}}{=}} \newcommand{\eqsp}{\;} \newcommand{\Eset}{ {\mathsf E}} \newcommand{\esssup}{\mathrm{essup}} \newcommand{\fr}[1]{{\left\langle #1 \right\rangle}} \newcommand{\falph}{f} \renewcommand{\geq}{\geqslant} \newcommand{\hchi}{\hat \chi} \newcommand{\Hset}{\mathsf{H}} \newcommand{\Id}{\mathrm{Id}} \newcommand{\img}{\text{Im}} \newcommand{\indi}[1]{\mathbf{1}_{#1}} \newcommand{\indiacc}[1]{\mathbf{1}_{\{#1\}}} \newcommand{\indin}[1]{\mathbf{1}\{#1\}} \newcommand{\itemm}{\quad \quad \blacktriangleright \;} \newcommand{\ker}{\text{Ker}} \newcommand{\klbck}[2]{\mathrm{K}\lr{#1||#2}} \newcommand{\law}{\mathcal{L}} \newcommand{\labelinit}{\pi} \newcommand{\labelkernel}{Q} \renewcommand{\leq}{\leqslant} \newcommand{\lone}{\mathsf{L}_1} \newcommand{\lrav}[1]{\left|#1 \right|} \newcommand{\lr}[1]{\left(#1 \right)} \newcommand{\lrb}[1]{\left[#1 \right]} \newcommand{\lrc}[1]{\left\{#1 \right\}} \newcommand{\lrcb}[1]{\left\{#1 \right\}} \newcommand{\ltwo}[1]{\PE^{1/2}\lrb{\lrcb{#1}^2}} \newcommand{\Ltwo}{\mathrm{L}^2} \newcommand{\mc}[1]{\mathcal{#1}} \newcommand{\mcbb}{\mathcal B} \newcommand{\mcf}{\mathcal{F}} \newcommand{\meas}[1]{\mathrm{M}_{#1}} \newcommand{\norm}[1]{\left\|#1\right\|} \newcommand{\normmat}[1]{{\left\vert\kern-0.25ex\left\vert\kern-0.25ex\left\vert #1 \right\vert\kern-0.25ex\right\vert\kern-0.25ex\right\vert}} \newcommand{\nset}{\mathbb N} \newcommand{\one}{\mathsf{1}} \newcommand{\PE}{\mathbb E} \newcommand{\PP}{\mathbb P} \newcommand{\projorth}[1]{\mathsf{P}^\perp_{#1}} \newcommand{\Psif}{\Psi_f} \newcommand{\pscal}[2]{\langle #1,#2\rangle} \newcommand{\pscal}[2]{\langle #1,#2\rangle} \newcommand{\psconv}{\stackrel{\PP-a.s.}{\rightarrow}} \newcommand{\qset}{\mathbb Q} \newcommand{\rmd}{\mathrm d} \newcommand{\rme}{\mathrm e} \newcommand{\rmi}{\mathrm i} \newcommand{\Rset}{\mathbb{R}} \newcommand{\rset}{\mathbb{R}} \newcommand{\rti}{\sigma} \newcommand{\section}[1]{==== #1 ====} \newcommand{\seq}[2]{\lrc{#1\eqsp: \eqsp #2}} \newcommand{\set}[2]{\lrc{#1\eqsp: \eqsp #2}} \newcommand{\sg}{\mathrm{sgn}} \newcommand{\supnorm}[1]{\left\|#1\right\|_{\infty}} \newcommand{\thv}{{\theta_\star}} \newcommand{\tmu}{ {\tilde{\mu}}} \newcommand{\Tset}{ {\mathsf{T}}} \newcommand{\Tsigma}{ {\mathcal{T}}} \newcommand{\ttheta}{{\tilde \theta}} \newcommand{\tv}[1]{\left\|#1\right\|_{\mathrm{TV}}} \newcommand{\unif}{\mathrm{Unif}} \newcommand{\weaklim}[1]{\stackrel{\mathcal{L}_{#1}}{\rightsquigarrow}} \newcommand{\Xset}{{\mathsf X}} \newcommand{\Xsigma}{\mathcal X} \newcommand{\Yset}{{\mathsf Y}} \newcommand{\Ysigma}{\mathcal Y} \newcommand{\Var}{\mathbb{V}\mathrm{ar}} \newcommand{\zset}{\mathbb{Z}} \newcommand{\Zset}{\mathsf{Z}} $$

2017/10/07 23:39 · douc

$$ \newcommand{\exempleOne}{Exemple} \newcommand{\exempleTwo}{Exemple} $$

2019/02/13 08:39 · douc

Optimization for machine/deep learning

Let $f: \mathbb{R}^d \to \mathbb{R}$ be a differentiable function. If $x^{\star}$ is a local extremum then $\nabla f(x^{\star}) = 0$, where $\nabla f(x^{\star})$ is the gradient of the function $f$ at point $x^{\star}$: $$ \nabla f(x) = \left( \frac{\partial f}{\partial x_1}, \ldots, \frac{\partial f}{\partial x_d}\right)^T\,. $$ For instance, if $f: x \mapsto \langle a;x\rangle$, then $\nabla f(x) = a$ and if $f: x \mapsto x^T A x$ then $\nabla f(x) = (A + A^T) x$. The gradient is orthogonal to level sets where for a function $f: \mathbb{R}^d \to \mathbb{R}$ and a value $c_\star\in\mathbb{R}$, the level set associated with $c_\star$ is given by $\mathcal{C}_{c_\star} = \{x \in \mathbb{R}^d, f(x) = c_\star\}$.

Deep learning optimization problems may be written as (see previous session for logistic regression and the multilayer perceptron): $$ \mathrm{argmin}_{w \in \mathbb{R}^d}\left\{\underbrace{\frac{1}{n}\sum_{i=1}^n \ell(Y_i, \langle w; X_i \rangle) + \lambda g(w)}_{f(w)}\right\}\,, $$ where $\ell$ is a given loss function and $g$ is a penalization function such as $g: w \mapsto \|w\|_2^2$ or $g: w \mapsto \|w\|_1^2$. The objective of gradient based optimization algorithms is to produce a sequence of parameter estimates $(w^{(k)})_{k\geqslant 0}$ which approximate this minimum.

Convergence of an elementary gradient descent algorithm

Starting from an initial vector $w^{(0)}$ and with a step size $\eta>0$, the most simple gradient based algorithm writes, for all $k\geqslant 0$, $$ w^{(k+1)} = w^{(k)} - \eta \nabla f(w^{(k)})\,. $$ Assume that $f : \mathbb{R}^d \to \mathbb{R}$ is a $L$-smooth convex function, i.e. a convex function such that for all $(x,y)\in\rset^d\times\rset^d$, $$ \|\nabla f(x) - \nabla f(y)\|\leqslant L\|x-y\| $$ and let $w^{\star}$ such that $f(w^\star)$ is the minimum of $f$ on $\mathbb{R}^d$. Then, Gradient Descent with step size $\eta \leqslant 1/L$ satisfies $$ f(w^{(k)}) - f(w^{\star}) \leqslant \frac{\|w^{(0)} - w^{\star}\|_2^2}{2 \eta k}\,. $$ In particular, for $\eta=1/L$, $$ L\|w^{(0)} - w^{\star}\|_2^2/2\varepsilon $$ iterations are sufficient to get an $\varepsilon$-approximation of the minimal value of $f$. The proof of this result relies on the descent lemma: if $f$ is $L$-smooth, then for any $w, w' \in \mathbb{R}^d$ $$ f(w') \leqslant f(w) + \langle \nabla f(w); w' - w\rangle + \frac {L}{2} \|w - w'\|_2^2\,. $$ Assuming this holds, remark that \begin{equation*} \mathrm{argmin}_{w \in \mathbb{R}^d}\left\{ f(w^{(k)} + \langle \nabla f(w^{(k)}; w - w^{(k)}\rangle + \frac{L}{2} \|w -w^{(k)}\|_2^2\right\} = \mathrm{argmin}_{w \in \mathbb{R}^d} \left\| w - (w^{(k)} - \frac 1 {L} \nabla f(w^{(k)})\right\|_2^2\,. \end{equation*} Hence, it is natural to choose $$ w^{(k+1)} = w^{(k)} - \frac 1 {L} \nabla f(w^{(k)})\,. $$

Faster rate for strongly convex function

A function $f: \mathbb{R}^d \to \mathbb{R}$ is $\mu$-strongly convex if $$ x \mapsto f(x) - \frac{\mu}{2}\|x\|_2^2 $$ is convex. If $f$ is differentiable it is equivalent to assume that, for all $x \in \mathbb{R}^d$, $$ \lambda_{min}(\nabla^2 f(x)) \geqslant \mu\,, $$ where $\lambda_{min}(\nabla^2 f(x))$ is the smallest eigenvalue of $\nabla^2 f(x)$ which is the Hessain matrix of $f$ at $x$ ($d\times d$ matrix containing second order derivatives of $f$ at $x$). This is also equivalent to assume that, for all $x,y \in \mathbb{R}^d$, $$ f(y) \geqslant f(x) + \langle \nabla f (x), y-x\rangle + \frac{\mu}{2} \|y-x\|_2^2. $$ If $f : \mathbb{R}^d \to \mathbb{R}$ is a $L$-smooth, $\mu$ strongly convex function and $w^{\star}$ is a minimum of $f$ on $\mathbb{R}^d$, then, Gradient Descent with step size $\eta \leqslant 1/L$ satisfies $$ f(w^{(k)}) - f(w^{\star}) \leq \Big( 1 - \eta \mu \Big)^k \|f(w^{(0)}) - f(w^{\star})\|_2^2. $$

Choice of $\eta$

In practice, the algorithm performance is highly sensitive to the choice of $\eta$.

(Exact line search). At each step, choose the best $\eta$ by optimizing $$ \eta^{(k)} = \argmin_{\eta >0} f(w - \eta \nabla f(w))\,. $$ The computational cost of this approach makes it prohibitive.

(Backtracking line search). First, fix a parameter $0 < \beta < 1$, then at each iteration, start with $\eta^{(k)}=1$ and while $$ f( w^{(k)} - \eta^{(k)}\nabla f (w^{(k)})) > f(w^{(k)}) - \frac{\eta^{(k)}}{2} \|\nabla f(w^{(k)}) \|^2, $$ update $ \eta^{(k)} = \beta \eta^{(k)}$. This simple to implement approach is widespread and efficient in practice. Let $f : \mathbb{R}^d \to \mathbb{R}$ be a $L$-smooth convex function and $w^{\star}$ be a minimum of $f$ on $\mathbb{R}^d$. Then, Gradient Descent with backtracking line search satisfies $$ f(w^{(k)}) - f(w^{\star}) \leqslant \frac{\|w^{(0)} - w^{\star}\|_2^2}{2 k \min(1, \beta/L) }\,, $$ which provides a convergence result for a method with adaptive step size (i.e. it is not required to choose and tune $\eta$).

Stochastic gradient descent

In the case of machine/deep learning applications, the function to be minimized is of the form: $$ f:w\mapsto \frac{1}{n}\sum_{i=1}^n \ell(Y_i, \langle w; X_i \rangle) + \lambda g(w) = \frac{1}{n}\sum_{i=1}^n f_i(w)\,. $$ Previous methods are based on \textbf{full gradients}, since at each iteration they require to compute $$ \nabla f(w) = \frac 1n \sum_{i=1}^n \nabla f_i(w)\,, $$ which depends on the whole dataset. When processing very large datasets ($n$ is large), this approach has a highly prohibitive computational cost for a {\bf unique step} towards the minimum. Stochastic Gradient Descent algorithm aims at reducing this computational cost by replacing this gradient by an unbiased estimate. If $I$ choose uniformly at random $I \in \{ 1, \ldots, n \}$, then $$ \mathbb{E}[\nabla f_I(w)] = \frac 1n \sum_{i=1}^n \nabla f_{i}(w) = \nabla f(w)\,. $$ Therefore $\nabla f_I(w)$ is an \textbf{unbiased} estimate of the full gradient $\nabla f(w)$. The computation of $\nabla f_I(w)$ only requires the $I$-th data and its complexity is of order $O(d)$. Starting from $w^{(0)}$, and chosing the step size $\eta_k$ at each time step, Stochastic Gradient Descent algorithm proceeds as follows. For all $k\geqslant 1$, pick at random (uniformly) $I_k$ in $\{ 1, \ldots, n \}$ and set $$ w^{(k)} = w^{(k-1)} - \eta_k \nabla f_{I_k}(w^{(k-1)})\,. $$ Each iteration has complexity $O(d)$ instead of $O(nd)$ for the previous full gradient methods. Consider the projected stochastic gradient descent algorithm on the ball $B(0,R)$ with $R>0$ fixed. Assume that $f$ is convex and that there exists $M>0$ satisfying, for all $x \in B(0,R)$, $$ \|\nabla f_(x)\| \leqslant M $$ and assume that all minima of $f$ belong to $B(0,R)$. Then, setting $\eta_k = 2R/(M\sqrt{k})$, $$ \mathbb{E}\bigg[ f\Big( \frac{1}{k} \sum_{j = 1}^k w^{(j)}\Big)\bigg] - f(w^{\star}) \leqslant \frac{3RM}{\sqrt{k}}\,. $$ In the case where $f$ is $\mu$ strongly convex, then, setting $\eta_k = 2/(\mu(k+1))$ yields $$ \mathbb{E}\Big[ f\Big( \frac{2}{k(k+1)} \sum_{j = 1}^k j w^{(j-1)}\Big)\Big] - f(w^{\star}) \leqslant \frac{2b^2}{\mu(k+1)}\,. $$ Stochastic Gradient Descent performance may be improved using variance reduction techniques, see for instance Stochastic Variance Reduced Gradient (SVRG) (Johnson et al. 2013), Stochastic Average Gradient (SAG) (Schmidt et al., 2017) and SAGA (Defazio et al., 2014).

Nesterov Accelerated Gradient Descent

Momentum based gradient descent algorithms aim at taking into account the previous updates as additional velocity to avoid convergence to local minima. Starting from $w^{(0)}$ and an initial velocity $v^{(0)}=0$ and using a sequence of step sizes $(\eta_k)_{k\geqslant 0}$ and momentum $(\beta_k)_{k\geqslant 0}$ in $(0,1)$, the Nesterov Accelerated Gradient Descent proceeds as follows: \begin{align*} v^{(k+1)} &= w^{(k)} - \eta \nabla f (w^{(k)})\,,\\ w^{(k+1)} &= v^{(k+1)} + \beta_{k+1}( v^{(k+1)} - v^{(k)})\,. \end{align*} If the target function $f$ is a $L$-smooth, convex and if $w^{\star}$ is a minimum of $f$ then, if $\beta_{k+1} = k/(k+3)$, $$ f(w^{(k)}) - f(w^{\star}) \leqslant \frac{2 \|w^{(0)} - w^{\star}\|_2^2}{\eta (k+1)^2}\,. $$ When in addition $f$ is $\mu$ strongly convex, choosing $$ \beta_{k} = \frac{1 - \sqrt{\mu/L}}{1 + \sqrt{\mu/L}}\,, $$ yields $$ f(w^{(k)}) - f(w^{\star}) \leqslant \frac{ \|w^{(0)} - w^{\star}\|_2^2}{\eta} \Big( 1 - \sqrt{\frac{\mu}{L}}\Big)^k\,. $$ It can be proved that the Nesterov Accelerated Gradient Descent reaches the optimal bounds among optimization algorithms such that for all $k\geqslant 0$, $$ w^{(k)} \in w^{(0)} + \textrm{Span}(\nabla f (w^{(0)}), \ldots, \nabla f (w^{(k-1)})). $$ This means that it is possible to design a $L$-smooth convex function such that any algorithm of this kind has an error term greater than $(k+1)^{-2}$ up to a multiplicative constant after $k$ step of the algorithm.

Optimization methods for neural networks

(AdaGrad). The ADAptive GRADient algorithm introduced by (Duchi et al. 2011) starts from $w^{(0)}$ and uses a learning rate $\eta > 0$ and a momentum $\alpha$ and defines, for all $k\geqslant 0$ and all $j \in\{1, \ldots , d\}$, $$ w_j^{(k+1)} \gets w_j^{(k)} - \frac{\eta}{\sqrt{\sum_{\tau=1}^k (\nabla f (w^{(\tau)}))_j^2}} (\nabla f(w^{(k)}))_j\,. $$ The rationale of this method is that different rates are used for all coordinates which is crucial for neural networks in which gradient at different layers can be of different order of magnitude. It is proved in |(Ward et al., 2018) that AdaGrad achieves the same convergence rate as gradient descent with optimal fixed stepsize up to a $\log$ factor. The adaptive step size grows with the inverse of the gradient magnitudes, so that large gradients have small learning rates and small gradients have large learning rates.

(AdaDelta). Introduced in (Zeiler, 2012), was introduced to reduce the sensitivity to initial conditions of AdaGrad. Indeed, if the initial gradients are large, the learning rates of AdaGrad will be low for all updates which can be overcome by increasing $\eta$, but making the AdaGrad method highly sensitive to the choice of $\eta$.

(RMSprop optimizer). Unpublished method, from the course of |Geoff Hinton.

(ADAM: Adaptive moment estimation). Introduced in (Kingma et al., 2014) and considered as the state of the art to otpimize neural networks, the ADAM procedure update the parameter estimate as follows. Starting from $m_0 = 0$ and $v_0=0$ and choosing $\beta_1, \beta_2, \eta, \varepsilon\in (0,1)$, compute first and second moment estimate $$ m_k = \beta_1 m_{k-1} + (1 - \beta_1) \nabla f (w^{(k)}) \quad \mathrm{and} \quad v_k = \beta_2 v_{k-1} + (1 - \beta_2) (\nabla f (w^{(k)}))^2\,, $$ then, compute the correction terms $$ \hat{m}_k = \frac{m_k}{1 - \beta_1^k} \quad \hat{v}_k = \frac{v_k}{1 - \beta_2^k}\,, $$ and update the parameter estimate with $$ w^{(k+1)} = w^{(k)} - \frac{\eta}{\sqrt{\hat{v}_k}+ \varepsilon} \hat{m}_k\,. $$ First convergence results can be found in |(Kingma et al., 2014) and examples where ADAM algorithm does not converge to the optimum are given in (Reddi et al., 2018). Recent analysis by (Barakat et al., 2018).

world/datascience/optimization.txt · Last modified: 2022/03/16 07:40 (external edit)