• Login
    View Item 
    •   DSpace@RPI Home
    • Rensselaer Libraries
    • z_[technical processing]
    • View Item
    •   DSpace@RPI Home
    • Rensselaer Libraries
    • z_[technical processing]
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Accelerated stochastic gradient methods with adaptive techniques and distributed computing

    Author
    Yan, Yonggui
    View/Open
    Yan_rpi_0185E_12184.pdf (6.503Mb)
    Other Contributors
    Xu, Yangyang, Y; Lai, Rongjie, R; Mitchell, John, J; Ji, Qiang, Q;
    Date Issued
    2023-05
    Subject
    Mathematics
    Degree
    PhD;
    Terms of Use
    This electronic version is a licensed copy owned by Rensselaer Polytechnic Institute (RPI), Troy, NY. Copyright of original work retained by author.;
    Metadata
    Show full item record
    URI
    https://hdl.handle.net/20.500.13015/6650
    Abstract
    Stochastic gradient methods (SGMs) have gained widespread usage for solving stochastic optimization problems due to their simplicity and efficient computation as first-order methods. However, vanilla SGMs suffer from slow convergence, prompting the development of many adaptive variants to expedite convergence. With the exponential growth of data, it has become increasingly challenging to process all the data on a single machine within a reasonable amount of time. To address this challenge, leveraging the power of multiple machines in parallel has become an affordable and effective solution to reduce computing time. Despite the popularity of adaptive techniques and parallelization for large-scale data processing, the analysis of adaptive SGMs and distributed methods is often restricted to problems with no constraints or easy-to-project constraints, convex problems, or nonconvex but smooth problems. Many applications are in uncovered forms and remain under-explored, such as Neyman-Pearson classification, fairness-constrained classification, phase retrieval, and sparsity-regularized deep learning. To address these challenges, my research aims to accelerate SGMs by using adaptive techniques in uncharted situations and modify the methods for distributed settings. I have proposed three methods. The first method solves expectation-constrained convex stochastic programs by an accelerated primal-dual SGM. The second method tackles nonconvex (and possibly nonsmooth) programs by an accelerated SGM in a centralized distributed system where derivatives are computed on stale variables. The third method focuses on solving nonconvex stochastic composite problems in decentralized distributed systems with heterogeneous data distributions. These methods have the potential to expand the application of SGMs to a wider range of challenging problems and save the computation time for large-scale datasets.;
    Description
    May2023; School of Science
    Department
    Dept. of Mathematical Sciences;
    Publisher
    Rensselaer Polytechnic Institute, Troy, NY
    Relationships
    Rensselaer Theses and Dissertations Online Collection;
    Access
    Users may download and share copies with attribution in accordance with a Creative Commons Attribution-Noncommercial-No Derivative Works 3.0 license. No commercial use or derivatives are permitted without the explicit approval of the author.;
    Collections
    • z_[technical processing]

    Browse

    All of DSpace@RPICommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    Login

    DSpace software copyright © 2002-2023  DuraSpace
    Contact Us | Send Feedback
    DSpace Express is a service operated by 
    Atmire NV