On variance reduction for stochastic smooth convex optimization with multiplicative noise

On variance reduction for stochastic smooth convex optimization with multiplicative noise Math. Program., Ser. B https://doi.org/10.1007/s10107-018-1297-x FULL LENGTH PAPER On variance reduction for stochastic smooth convex optimization with multiplicative noise 1 2 Alejandro Jofré · Philip Thompson Received: 24 May 2017 / Accepted: 11 May 2018 © Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society 2018 Abstract We propose dynamic sampled stochastic approximation (SA) methods for stochastic optimization with a heavy-tailed distribution (with finite 2nd moment). The objective is the sum of a smooth convex function with a convex regularizer. Typically, it is assumed an oracle with an upper bound σ on its variance (OUBV). Differently, we assume an oracle with multiplicative noise. This rarely addressed setup is more aggressive but realistic, where the variance may not be uniformly bounded. Our methods achieve optimal iteration complexity and (near) optimal oracle complexity. For the smooth convex class, we use an accelerated SA method a la FISTA which achieves, given tolerance ε> 0, the optimal iteration complexity of O(ε ) with a −2 − 2 near-optimal oracle complexity of O(ε )[ln(ε )] . This improves upon Ghadimi and Lan (Math Program 156:59–99, 2016)whereitisassumedanOUBV. Forthe −1 strongly convex class, our method achieves optimal iteration complexity of O(ln(ε )) −1 and http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Mathematical Programming Springer Journals

On variance reduction for stochastic smooth convex optimization with multiplicative noise

Loading next page...
 
/lp/springer_journal/on-variance-reduction-for-stochastic-smooth-convex-optimization-with-BfoPvwtEnP
Publisher
Springer Berlin Heidelberg
Copyright
Copyright © 2018 by Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society
Subject
Mathematics; Calculus of Variations and Optimal Control; Optimization; Mathematics of Computing; Numerical Analysis; Combinatorics; Theoretical, Mathematical and Computational Physics; Mathematical Methods in Physics
ISSN
0025-5610
eISSN
1436-4646
D.O.I.
10.1007/s10107-018-1297-x
Publisher site
See Article on Publisher Site

Abstract

Math. Program., Ser. B https://doi.org/10.1007/s10107-018-1297-x FULL LENGTH PAPER On variance reduction for stochastic smooth convex optimization with multiplicative noise 1 2 Alejandro Jofré · Philip Thompson Received: 24 May 2017 / Accepted: 11 May 2018 © Springer-Verlag GmbH Germany, part of Springer Nature and Mathematical Optimization Society 2018 Abstract We propose dynamic sampled stochastic approximation (SA) methods for stochastic optimization with a heavy-tailed distribution (with finite 2nd moment). The objective is the sum of a smooth convex function with a convex regularizer. Typically, it is assumed an oracle with an upper bound σ on its variance (OUBV). Differently, we assume an oracle with multiplicative noise. This rarely addressed setup is more aggressive but realistic, where the variance may not be uniformly bounded. Our methods achieve optimal iteration complexity and (near) optimal oracle complexity. For the smooth convex class, we use an accelerated SA method a la FISTA which achieves, given tolerance ε> 0, the optimal iteration complexity of O(ε ) with a −2 − 2 near-optimal oracle complexity of O(ε )[ln(ε )] . This improves upon Ghadimi and Lan (Math Program 156:59–99, 2016)whereitisassumedanOUBV. Forthe −1 strongly convex class, our method achieves optimal iteration complexity of O(ln(ε )) −1 and

Journal

Mathematical ProgrammingSpringer Journals

Published: Jun 5, 2018

References

You’re reading a free preview. Subscribe to read the entire article.


DeepDyve is your
personal research library

It’s your single place to instantly
discover and read the research
that matters to you.

Enjoy affordable access to
over 18 million articles from more than
15,000 peer-reviewed journals.

All for just $49/month

Explore the DeepDyve Library

Search

Query the DeepDyve database, plus search all of PubMed and Google Scholar seamlessly

Organize

Save any article or search result from DeepDyve, PubMed, and Google Scholar... all in one place.

Access

Get unlimited, online access to over 18 million full-text articles from more than 15,000 scientific journals.

Your journals are on DeepDyve

Read from thousands of the leading scholarly journals from SpringerNature, Elsevier, Wiley-Blackwell, Oxford University Press and more.

All the latest content is available, no embargo periods.

See the journals in your area

DeepDyve

Freelancer

DeepDyve

Pro

Price

FREE

$49/month
$360/year

Save searches from
Google Scholar,
PubMed

Create lists to
organize your research

Export lists, citations

Read DeepDyve articles

Abstract access only

Unlimited access to over
18 million full-text articles

Print

20 pages / month

PDF Discount

20% off