Accelerated stochastic gradient methods with adaptive techniques and distributed computing
Loading...
Authors
Yan, Yonggui
Issue Date
2023-05
Type
Electronic thesis
Thesis
Thesis
Language
en_US
Keywords
Mathematics
Alternative Title
Abstract
Stochastic gradient methods (SGMs) have gained widespread usage for solving stochastic optimization problems due to their simplicity and efficient computation as first-order methods. However, vanilla SGMs suffer from slow convergence, prompting the development of many adaptive variants to expedite convergence. With the exponential growth of data, it has become increasingly challenging to process all the data on a single machine within a reasonable amount of time. To address this challenge, leveraging the power of multiple machines in parallel has become an affordable and effective solution to reduce computing time. Despite the popularity of adaptive techniques and parallelization for large-scale data processing, the analysis of adaptive SGMs and distributed methods is often restricted to problems with no constraints or easy-to-project constraints, convex problems, or nonconvex but smooth problems. Many applications are in uncovered forms and remain under-explored, such as Neyman-Pearson classification, fairness-constrained classification, phase retrieval, and sparsity-regularized deep learning. To address these challenges, my research aims to accelerate SGMs by using adaptive techniques in uncharted situations and modify the methods for distributed settings. I have proposed three methods. The first method solves expectation-constrained convex stochastic programs by an accelerated primal-dual SGM. The second method tackles nonconvex (and possibly nonsmooth) programs by an accelerated SGM in a centralized distributed system where derivatives are computed on stale variables. The third method focuses on solving nonconvex stochastic composite problems in decentralized distributed systems with heterogeneous data distributions. These methods have the potential to expand the application of SGMs to a wider range of challenging problems and save the computation time for large-scale datasets.
Description
May2023
School of Science
School of Science
Full Citation
Publisher
Rensselaer Polytechnic Institute, Troy, NY