Taming Convergence for Asynchronous Stochastic Gradient Descent with Unbounded Delay in Non-Convex Learning

dc.contributor.author Zhang, Xin
dc.contributor.author Zhu, Zhengyuan
dc.contributor.author Liu, Jia
dc.contributor.author Zhu, Zhengyuan
dc.contributor.department Statistics
dc.date 2019-06-10T09:29:24.000
dc.date.accessioned 2020-07-02T06:56:01Z
dc.date.available 2020-07-02T06:56:01Z
dc.date.issued 2018-05-24
dc.description.abstract <p>Understanding the convergence performance of asynchronous stochastic gradient descent method (Async-SGD) has received increasing attention in recent years due to their foundational role in machine learning. To date, however, most of the existing works are restricted to either bounded gradient delays or convex settings. In this paper, we focus on Async-SGD and its variant Async-SGDI (which uses increasing batch size) for non-convex optimization problems with unbounded gradient delays. We prove o(1/k−−√)convergence rate for Async-SGD and o(1/k) for Async-SGDI. Also, a unifying sufficient condition for Async-SGD's convergence is established, which includes two major gradient delay models in the literature as special cases and yields a new delay model not considered thus far.</p>
dc.description.comments <p>This pre-print is made available through arXiv: <a href="https://arxiv.org/abs/1805.09470">https://arxiv.org/abs/1805.09470</a>.</p>
dc.identifier archive/lib.dr.iastate.edu/stat_las_preprints/142/
dc.identifier.articleid 1141
dc.identifier.contextkey 14319722
dc.identifier.s3bucket isulib-bepress-aws-west
dc.identifier.submissionpath stat_las_preprints/142
dc.identifier.uri https://dr.lib.iastate.edu/handle/20.500.12876/90303
dc.language.iso en
dc.source.bitstream archive/lib.dr.iastate.edu/stat_las_preprints/142/2018_Zhu_TamingConvergencePreprint.pdf|||Fri Jan 14 20:16:13 UTC 2022
dc.subject.disciplines Computer Sciences
dc.subject.disciplines Statistics and Probability
dc.subject.keywords Asynchronous stochastic gradient descent
dc.subject.keywords parallel distributed computing
dc.subject.keywords convergence rate
dc.subject.keywords unbounded delay
dc.title Taming Convergence for Asynchronous Stochastic Gradient Descent with Unbounded Delay in Non-Convex Learning
dc.type article
dc.type.genre article
dspace.entity.type Publication
relation.isAuthorOfPublication 51db2a08-8f9d-4f97-bdbc-6790b3d5a608
relation.isOrgUnitOfPublication 264904d9-9e66-4169-8e11-034e537ddbca
File
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
2018_Zhu_TamingConvergencePreprint.pdf
Size:
748.14 KB
Format:
Adobe Portable Document Format
Description:
Collections