Accelerate stochastic subgradient method by leveraging local growth condition

Author:

Xu Yi1,Lin Qihang2,Yang Tianbao1

Affiliation:

1. Department of Computer Science, The University of Iowa, Iowa City, IA 52242, USA

2. Department of Management Sciences, The University of Iowa, Iowa City, IA 52242, USA

Abstract

In this paper, a new theory is developed for first-order stochastic convex optimization, showing that the global convergence rate is sufficiently quantified by a local growth rate of the objective function in a neighborhood of the optimal solutions. In particular, if the objective function [Formula: see text] in the [Formula: see text]-sublevel set grows as fast as [Formula: see text], where [Formula: see text] represents the closest optimal solution to [Formula: see text] and [Formula: see text] quantifies the local growth rate, the iteration complexity of first-order stochastic optimization for achieving an [Formula: see text]-optimal solution can be [Formula: see text], which is optimal at most up to a logarithmic factor. To achieve the faster global convergence, we develop two different accelerated stochastic subgradient methods by iteratively solving the original problem approximately in a local region around a historical solution with the size of the local region gradually decreasing as the solution approaches the optimal set. Besides the theoretical improvements, this work also includes new contributions toward making the proposed algorithms practical: (i) we present practical variants of accelerated stochastic subgradient methods that can run without the knowledge of multiplicative growth constant and even the growth rate [Formula: see text]; (ii) we consider a broad family of problems in machine learning to demonstrate that the proposed algorithms enjoy faster convergence than traditional stochastic subgradient method. We also characterize the complexity of the proposed algorithms for ensuring the gradient is small without the smoothness assumption.

Publisher

World Scientific Pub Co Pte Lt

Subject

Applied Mathematics,Analysis

Cited by 4 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Stochastic Localization Methods for Convex Discrete Optimization via Simulation;Operations Research;2023-10-06

2. Snacks: a fast large-scale kernel SVM solver;2023 European Control Conference (ECC);2023-06-13

3. Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization;SIAM Journal on Mathematics of Data Science;2022-05-12

4. On the Adaptivity of Stochastic Gradient-Based Optimization;SIAM Journal on Optimization;2020-01

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3