Show simple item record

dc.contributor.authorKarimi, Sahar 20:23:30 (GMT) 20:23:30 (GMT)
dc.description.abstractIn a series of work initiated by Nemirovsky and Yudin, and later extended by Nesterov, first-order algorithms for unconstrained minimization with optimal theoretical complexity bound have been proposed. On the other hand, conjugate gradient algorithms as one of the widely used first-order techniques suffer from the lack of a finite complexity bound. In fact their performance can possibly be quite poor. This dissertation is partially on tightening the gap between these two classes of algorithms, namely the traditional conjugate gradient methods and optimal first-order techniques. We derive conditions under which conjugate gradient methods attain the same complexity bound as in Nemirovsky-Yudin's and Nesterov's methods. Moreover, we propose a conjugate gradient-type algorithm named CGSO, for Conjugate Gradient with Subspace Optimization, achieving the optimal complexity bound with the payoff of a little extra computational cost. We extend the theory of CGSO to convex problems with linear constraints. In particular we focus on solving $l_1$-regularized least square problem, often referred to as Basis Pursuit Denoising (BPDN) problem in the optimization community. BPDN arises in many practical fields including sparse signal recovery, machine learning, and statistics. Solving BPDN is fairly challenging because the size of the involved signals can be quite large; therefore first order methods are of particular interest for these problems. We propose a quasi-Newton proximal method for solving BPDN. Our numerical results suggest that our technique is computationally effective, and can compete favourably with the other state-of-the-art solvers.en
dc.publisherUniversity of Waterlooen
dc.subjectConvex Optimizationen
dc.subjectFirst-Order Methodsen
dc.subjectConjugate Gradienten
dc.subjectProximal Quasi-Newton Methodsen
dc.subjectBasis Pursuit Denoising Problemen
dc.subjectL1-regularized Least Square problemen
dc.titleOn the Relationship between Conjugate Gradient and Optimal First-Order Methods for Convex Optimizationen
dc.typeDoctoral Thesisen
dc.subject.programCombinatorics and Optimizationen and Optimizationen
uws-etd.degreeDoctor of Philosophyen

Files in this item


This item appears in the following Collection(s)

Show simple item record


University of Waterloo Library
200 University Avenue West
Waterloo, Ontario, Canada N2L 3G1
519 888 4883

All items in UWSpace are protected by copyright, with all rights reserved.

DSpace software

Service outages