Download presentation

Presentation is loading. Please wait.

Published byTyrese Cavill Modified over 2 years ago

1
Semi-Stochastic Gradient Descent Methods Jakub Konečný University of Edinburgh BASP Frontiers Workshop January 28, 2014

2
Based on Basic Method: S2GD Konečný and Richtárik. Semi-Stochastic Gradient Descent Methods, December 2013 Mini-batching (& proximal setting): mS2GD Konečný, Liu, Richtárik and Takáč. mS2GD: mS2GD: Minibatch semi-stochastic gradient descent in the proximal setting, October 2014 Coordinate descent variant: S2CD Konečný, Qu and Richtárik. Semi-stochastic coordinate descent, December 2014

3
Introduction

4
Large scale problem setting Problems are often structured Frequently arising in machine learning Structure – sum of functions is BIG

5
Examples Linear regression (least squares) Logistic regression (classification)

6
Assumptions Lipschitz continuity of derivative of Strong convexity of

7
Gradient Descent (GD) Update rule Fast convergence rate Alternatively, for accuracy we need iterations Complexity of single iteration – (measured in gradient evaluations)

8
Stochastic Gradient Descent (SGD) Update rule Why it works Slow convergence Complexity of single iteration – (measured in gradient evaluations) a step-size parameter

9
Goal GD SGD Fast convergence gradient evaluations in each iteration Slow convergence Complexity of iteration independent of Combine in a single algorithm

10
Semi-Stochastic Gradient Descent S2GD

11
Intuition The gradient does not change drastically We could reuse the information from “old” gradient

12
Modifying “old” gradient Imagine someone gives us a “good” point and Gradient at point, near, can be expressed as Approximation of the gradient Already computed gradientGradient change We can try to estimate

13
The S2GD Algorithm Simplification; size of the inner loop is random, following a geometric rule

14
Theorem

15
Convergence rate How to set the parameters ? Can be made arbitrarily small, by decreasing For any fixed, can be made arbitrarily small by increasing

16
Setting the parameters The accuracy is achieved by setting Total complexity (in gradient evaluations) # of epochs full gradient evaluation cheap iterations # of epochs stepsize # of iterations Fix target accuracy

17
Complexity S2GD complexity GD complexity iterations complexity of a single iteration Total

18
Related Methods SAG – Stochastic Average Gradient (Mark Schmidt, Nicolas Le Roux, Francis Bach, 2013) Refresh single stochastic gradient in each iteration Need to store gradients. Similar convergence rate Cumbersome analysis SAGA (Aaron Defazio, Francis Bach, Simon Lacoste-Julien, 2014) Refined analysis MISO - Minimization by Incremental Surrogate Optimization (Julien Mairal, 2014) Similar to SAG, more general Elegant analysis, not really applicable for sparse data Julien is here

19
Related Methods SVRG – Stochastic Variance Reduced Gradient (Rie Johnson, Tong Zhang, 2013) Arises as a special case in S2GD Prox-SVRG (Tong Zhang, Lin Xiao, 2014) Extended to proximal setting EMGD – Epoch Mixed Gradient Descent (Lijun Zhang, Mehrdad Mahdavi, Rong Jin, 2013) Handles simple constraints, Worse convergence rate

20
Experiment (logistic regression on: ijcnn, rcv, real-sim, url)

21
Extensions

22
Sparse data For linear/logistic regression, gradient copies sparsity pattern of example. But the update direction is fully dense Can we do something about it? DENSESPARSE

23
Sparse data Yes we can! To compute, we only need coordinates of corresponding to nonzero elements of For each coordinate, remember when was it updated last time – Before computing in inner iteration number, update required coordinates Step being Compute direction and make a single update Number of iterations when the coordinate was not updated The “old gradient”

24
Sparse data implementation

25
High Probability Result The result holds only in expectation Can we say anything about the concentration of the result in practice? For any we have: Paying just logarithm of probability Independent from other parameters

26
Code Efficient implementation for logistic regression - available at MLOSS http://mloss.org/software/view/556/

27
mS2GD (mini-batch S2GD) How does mini-batching influence the algorithm? Replace by Provides two-fold speedup Provably less gradient evaluations are needed (up to certain number of mini-batches) Easy possibility of parallelism

28
S2CD (Semi-Stochastic Coordinate Descent) SGD type methods Sampling rows (training examples) of data matrix Coordinate Descent type methods Sampling columns (features) of data matrix Question: Can we do both? Sample both columns and rows

29
S2CD (Semi-Stochastic Coordinate Descent) Comlpexity S2GD

Similar presentations

OK

Lecture 29: Optimization and Neural Nets CS4670/5670: Computer Vision Kavita Bala Slides from Andrej Karpathy and Fei-Fei Li

Lecture 29: Optimization and Neural Nets CS4670/5670: Computer Vision Kavita Bala Slides from Andrej Karpathy and Fei-Fei Li

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Emergency dentist appt on your birthday Ppt on computer system for class 9 Ppt on product specification definition Ppt on area of parallelogram and triangles worksheets Ppt on regional transport office kerala Download ppt on conservation of fossil fuels Ppt on m-commerce and f-commerce Ppt on series and parallel circuits lessons Seminar ppt on power electronics Ppt on zener diode current