dc.rights.license | Users may download and share copies with attribution in accordance with a Creative Commons
Attribution-Noncommercial-No Derivative Works 3.0 license. No commercial use or derivatives
are permitted without the explicit approval of the author. | |
dc.contributor | Lai, Rongjie, R | |
dc.contributor | Mitchell, John, J | |
dc.contributor | Ji, Qiang, Q | |
dc.contributor.advisor | Xu, Yangyang, Y | |
dc.contributor.author | Yan, Yonggui | |
dc.date.accessioned | 2023-06-01T19:14:21Z | |
dc.date.available | 2023-06-01T19:14:21Z | |
dc.date.issued | 2023-05 | |
dc.identifier.uri | https://hdl.handle.net/20.500.13015/6650 | |
dc.description | May2023 | |
dc.description | School of Science | |
dc.description.abstract | Stochastic gradient methods (SGMs) have gained widespread usage for solving stochastic optimization problems due to their simplicity and efficient computation as first-order methods. However, vanilla SGMs suffer from slow convergence, prompting the development of many adaptive variants to expedite convergence. With the exponential growth of data, it has become increasingly challenging to process all the data on a single machine within a reasonable amount of time. To address this challenge, leveraging the power of multiple machines in parallel has become an affordable and effective solution to reduce computing time. Despite the popularity of adaptive techniques and parallelization for large-scale data processing, the analysis of adaptive SGMs and distributed methods is often restricted to problems with no constraints or easy-to-project constraints, convex problems, or nonconvex but smooth problems. Many applications are in uncovered forms and remain under-explored, such as Neyman-Pearson classification, fairness-constrained classification, phase retrieval, and sparsity-regularized deep learning. To address these challenges, my research aims to accelerate SGMs by using adaptive techniques in uncharted situations and modify the methods for distributed settings. I have proposed three methods. The first method solves expectation-constrained convex stochastic programs by an accelerated primal-dual SGM. The second method tackles nonconvex (and possibly nonsmooth) programs by an accelerated SGM in a centralized distributed system where derivatives are computed on stale variables. The third method focuses on solving nonconvex stochastic composite problems in decentralized distributed systems with heterogeneous data distributions. These methods have the potential to expand the application of SGMs to a wider range of challenging problems and save the computation time for large-scale datasets. | |
dc.language | ENG | |
dc.language.iso | en_US | |
dc.publisher | Rensselaer Polytechnic Institute, Troy, NY | |
dc.relation.ispartof | Rensselaer Theses and Dissertations Online Collection | |
dc.subject | Mathematics | |
dc.title | Accelerated stochastic gradient methods with adaptive techniques and distributed computing | |
dc.type | Electronic thesis | |
dc.type | Thesis | |
dc.date.updated | 2023-06-01T19:14:23Z | |
dc.rights.holder | This electronic version is a licensed copy owned by Rensselaer Polytechnic Institute (RPI), Troy, NY. Copyright of original work retained by author. | |
dc.description.degree | PhD | |
dc.relation.department | Dept. of Mathematical Sciences | |