Download presentation

Presentation is loading. Please wait.

Published byNasir Gane Modified over 2 years ago

1
Semi-Stochastic Gradient Descent Methods Jakub Konečný University of Edinburgh ETH Zurich November 3, 2014

2
Introduction

3
Large scale problem setting Problems are often structured Frequently arising in machine learning Structure – sum of functions is BIG

4
Examples Linear regression (least squares) Logistic regression (classification)

5
Assumptions Lipschitz continuity of derivative of Strong convexity of

6
Gradient Descent (GD) Update rule Fast convergence rate Alternatively, for accuracy we need iterations Complexity of single iteration – (measured in gradient evaluations)

7
Stochastic Gradient Descent (SGD) Update rule Why it works Slow convergence Complexity of single iteration – (measured in gradient evaluations) a step-size parameter

8
Goal GD SGD Fast convergence gradient evaluations in each iteration Slow convergence Complexity of iteration independent of Combine in a single algorithm

9
Semi-Stochastic Gradient Descent S2GD

10
Intuition The gradient does not change drastically We could reuse the information from “old” gradient

11
Modifying “old” gradient Imagine someone gives us a “good” point and Gradient at point, near, can be expressed as Approximation of the gradient Already computed gradientGradient change We can try to estimate

12
The S2GD Algorithm Simplification; size of the inner loop is random, following a geometric rule

13
Theorem

14
Convergence rate How to set the parameters ? Can be made arbitrarily small, by decreasing For any fixed, can be made arbitrarily small by increasing

15
Setting the parameters The accuracy is achieved by setting Total complexity (in gradient evaluations) # of epochs full gradient evaluation cheap iterations # of epochs stepsize # of iterations Fix target accuracy

16
Complexity S2GD complexity GD complexity iterations complexity of a single iteration Total

17
Related Methods SAG – Stochastic Average Gradient (Mark Schmidt, Nicolas Le Roux, Francis Bach, 2013) Refresh single stochastic gradient in each iteration Need to store gradients. Similar convergence rate Cumbersome analysis SAGA (Aaron Defazio, Francis Bach, Simon Lacoste-Julien, 2014) Refined analysis MISO - Minimization by Incremental Surrogate Optimization (Julien Mairal, 2014) Similar to SAG, slightly worse performance Elegant analysis

18
Related Methods SVRG – Stochastic Variance Reduced Gradient (Rie Johnson, Tong Zhang, 2013) Arises as a special case in S2GD Prox-SVRG (Tong Zhang, Lin Xiao, 2014) Extended to proximal setting EMGD – Epoch Mixed Gradient Descent (Lijun Zhang, Mehrdad Mahdavi, Rong Jin, 2013) Handles simple constraints, Worse convergence rate

19
Experiment (logistic regression on: ijcnn, rcv, real-sim, url)

20
Extensions

21
Sparse data For linear/logistic regression, gradient copies sparsity pattern of example. But the update direction is fully dense Can we do something about it? DENSESPARSE

22
Sparse data Yes we can! To compute, we only need coordinates of corresponding to nonzero elements of For each coordinate, remember when was it updated last time – Before computing in inner iteration number, update required coordinates Step being Compute direction and make a single update Number of iterations when the coordinate was not updated The “old gradient”

23
Sparse data implementation

24
S2GD+ Observing that SGD can make reasonable progress, while S2GD computes first full gradient (in case we are starting from arbitrary point), we can formulate the following algorithm (S2GD+)

25
S2GD+ Experiment

26
High Probability Result The result holds only in expectation Can we say anything about the concentration of the result in practice? For any we have: Paying just logarithm of probability Independent from other parameters

27
Code Efficient implementation for logistic regression - available at MLOSS http://mloss.org/software/view/556/

28
mS2GD (mini-batch S2GD) How does mini-batching influence the algorithm? Replace by Provides two-fold speedup Provably less gradient evaluations are needed (up to certain number of mini-batches) Easy possibility of parallelism Still preliminary work

29
S2CD (Semi-Stochastic Coordinate Descent) Coordinate updates? Sample non-uniformly and scale updates works Needs more cheaper iterations Still preliminary work

30
S2GD as a Learning Algorithm

31
Machine Learning Setting Space of input-output pairs Unknown distribution A relationship between inputs and outputs Loss function to measure discrepancy between predicted and real output Define Expected Risk

32
Machine Learning Setting Ideal goal: Find such that, But you cannot even evaluate Define Expected Risk

33
Machine Learning Setting We at least have iid samples Define Empirical Risk

34
First learning principle – fix a family of candidate prediction functions Find Empirical Minimizer Define Empirical Risk Machine Learning Setting

35
Since optimal is unlikely to belong to, we also define Define Empirical Risk Machine Learning Setting

36
Finding by minimizing the Empirical Risk exactly is often computationally expensive Run optimization algorithm that returns such that Define Empirical Risk Machine Learning Setting

37
Recapitulation Ideal optimum “Best” from our family Empirical Minimizer From approximate optimization

38
Machine Learning Goal Big goal is to minimize the Excess Risk Approximation error Estimation Error Optimization Error

39
Generic Machine Learning Problem All this leads to a complicated compromise Three variables Family of functions Number or examples Optimization accuracy Two constraints Maximal number of examples Maximal computational time available

40
Generic Machine Learning Problem Small scale learning problem If first inequality is tight Can reduce to insignificant levels and recover approximation-estimation tradeoff (well studied) Large scale learning problem If second inequality is tight More complicated compromise

41
Solving Large Scale ML Problem Several simplifications needed Not carefully balance the three terms; instead we only ensure that asymptotically Consider fixed family of functions, linearly parameterized by a vector Effectively setting to be a constant Simplifies to Estimation–Optimization tradeoff

42
Estimation–Optimization tradeoff Using uniform convergence bounds, one can obtain Often considered weak

43
Estimation–Optimization tradeoff Using Localized Bounds (Bousquet, PhD thesis, 2004) or Isomorphic Coordinate Projections (Bartlett and Mendelson, 2006), we get … if we can establish the following variance condition Often, for example under strong convexity, or making assumptions on the data distribution

44
Estimation–Optimization tradeoff Using the previous bounds yields where is an absolute constant We want to push this term below Choosing and using and we get the following table

Similar presentations

OK

Online Performance Guarantees for Sparse Recovery Raja Giryes ICASSP 2011 Volkan Cevher.

Online Performance Guarantees for Sparse Recovery Raja Giryes ICASSP 2011 Volkan Cevher.

© 2017 SlidePlayer.com Inc.

All rights reserved.

Ads by Google

Ppt on fuel cell Ppt on energy conservation Ppt on artificial intelligence download Ppt on 2d viewing pipeline Converter pub to ppt online maker Ppt on new technology in mechanical Thin film transistor display ppt on tv Led based moving message display ppt on ipad Ppt on different types of computer softwares types Ppt on mobile number portability in india