• Login
    View Item 
    •   DSpace@RPI Home
    • Rensselaer Libraries
    • RPI Theses Online (Complete)
    • View Item
    •   DSpace@RPI Home
    • Rensselaer Libraries
    • RPI Theses Online (Complete)
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Complementarity formulation for sparse optimization

    Author
    Tan, Jun
    View/Open
    Tan_rpi_0185E_11958.pdf (488.2Kb)
    Other Contributors
    Mitchell, John E.; Bennett, Kristin P.; Lai, Rongjie; Wang, Meng;
    Date Issued
    2021-12
    Subject
    Applied mathematics
    Degree
    PhD;
    Terms of Use
    This electronic version is a licensed copy owned by Rensselaer Polytechnic Institute (RPI), Troy, NY. Copyright of original work retained by author.;
    Metadata
    Show full item record
    URI
    https://hdl.handle.net/20.500.13015/6141
    Abstract
    Complementarity constraints are used to enforce the orthogonality between two decision vectors (or matrices). They naturally appear in mathematical programs for many applications and also in reformulations of certain non-convex optimization problems. Compared to a convex relaxation, a complementarity reformulation of a non-convex model can often yield a better-quality solution. In this thesis, I explore the complementarity reformulation to sparse optimization problems that can simultaneously include $\ell_0$ and rank minimization. Two applications are studied. One is the robust principal component analysis (RPCA), and the other is the low-rank representation (LRR) problem. For RPCA, I introduce complementarity constraints to both $\ell_0$-term and the rank function. For the LRR problem, I give complementarity reformulation to the original model and also a low-rank matrix decomposition based model. For the original model, the complementarity constraints are introduced to the rank function and also the column $\ell_0$-term, and for the low-rank decomposition model, the complementarity constraints are imposed to both low-rank factor matrices. Numerically, I apply the penalty method to each complementarity reformulated problem, with the block coordinate descent method for solving a sequence of penalty subproblems. Global (or whole-sequence) convergence and also local convergence (to global optimality) are shown. The effectiveness and efficiency of my methods are demonstrated by comparing to convex relaxation methods. In terms of recoverability, my method tends to give higher rate of successful recovery. Also, for the LRR problem, solving the low-rank matrix decomposition based model can save the computing time, as compared to solving the original model by the complementarity reformulation.;
    Description
    December 2021; School of Science
    Department
    Dept. of Mathematical Sciences;
    Publisher
    Rensselaer Polytechnic Institute, Troy, NY
    Relationships
    Rensselaer Theses and Dissertations Online Collection;
    Access
    Restricted to current Rensselaer faculty, staff and students in accordance with the Rensselaer Standard license. Access inquiries may be directed to the Rensselaer Libraries.;
    Collections
    • RPI Theses Online (Complete)

    Browse

    All of DSpace@RPICommunities & CollectionsBy Issue DateAuthorsTitlesSubjectsThis CollectionBy Issue DateAuthorsTitlesSubjects

    My Account

    Login

    DSpace software copyright © 2002-2023  DuraSpace
    Contact Us | Send Feedback
    DSpace Express is a service operated by 
    Atmire NV