Generalizations of the Alternating Direction Method of Multipliers for Large-Scale and Distributed Optimization

dc.contributor.advisorZhang, Yinen_US
dc.contributor.committeeMemberYin, Wotaoen_US
dc.contributor.committeeMemberJermaine, Christopheren_US
dc.contributor.committeeMemberTapia, Richarden_US
dc.creatorDeng, Weien_US
dc.date.accessioned2016-01-07T20:56:53Zen_US
dc.date.available2016-01-07T20:56:53Zen_US
dc.date.created2014-12en_US
dc.date.issued2014-11-19en_US
dc.date.submittedDecember 2014en_US
dc.date.updated2016-01-07T20:56:53Zen_US
dc.description.abstractDue to the dramatically increasing demand for dealing with "Big Data", efficient and scalable computational methods are highly desirable to cope with the size of the data. The alternating direction method of multipliers (ADMM), as a versatile algorithmic tool, has proven to be very effective at solving many large-scale and structured optimization problems, particularly arising from the areas of compressive sensing, signal and image processing, machine learning and applied statistics. Moreover, the algorithm can be implemented in a fully parallel and distributed manner to process huge datasets. These benefits have mainly contributed to the recent renaissance of ADMM for modern applications. This thesis makes important generalizations to ADMM to improve its flexibility and efficiency, as well as extending its convergence theory. Firstly, we allow more options of solving the subproblems either exactly or approximately, such as linearizing the subproblems, taking one gradient descent step, and approximating the Hessian. Often, when subproblems are expensive to solve exactly, it is much cheaper to compute approximate solutions to the subproblems which are still good enough to guarantee convergence. Although it may take more iterations to converge due to less accurate subproblems, the entire algorithm runs faster since each iteration takes much less time. Secondly, we establish the global convergence of these generalizations of ADMM. We further show the linear convergence rate under a variety of scenarios, which cover a wide range of applications in practice. Among these scenarios, we require that at least one of the two objective functions is strictly convex and has Lipschitz continuous gradient, along with certain full rank conditions on the constraint coefficient matrices. The derived rate of convergence also provides some theoretical guidance for optimizing the parameters of the algorithm. In addition, we introduce a simple technique to improve an existing convergence rate from O(1/k) to o(1/k). Thirdly, we introduce a parallel and multi-block extension to ADMM for solving convex separable problems with N blocks of variables. The algorithm decomposes the original problem into N smaller subproblems and solves them in parallel at each iteration. It is well suited to distributed computing and is particularly attractive for solving certain large-scale problems. We show that extending ADMM straightforwardly from the classic Gauss-Seidel setting to the Jacobi setting, from 2 blocks to N blocks, will preserve convergence if the constraint coefficient matrices are mutually near-orthogonal and have full column-rank. For general cases, we propose to add proximal terms of different kinds to the N subproblems so that they can be solved in flexible and efficient ways and the algorithm converges globally at a rate of o(1/k). We introduce a strategy for dynamically tuning the parameters of the algorithm, often leading to substantial acceleration of the convergence in practice. Numerical results are presented to demonstrate the efficiency of the proposed algorithm in comparison with several existing parallel algorithms. We also implemented our algorithm on Amazon EC2, an on-demand public computing cloud, and report its performance on very large-scale basis pursuit problems with distributed data.en_US
dc.format.mimetypeapplication/pdfen_US
dc.identifier.citationDeng, Wei. "Generalizations of the Alternating Direction Method of Multipliers for Large-Scale and Distributed Optimization." (2014) Diss., Rice University. <a href="https://hdl.handle.net/1911/87774">https://hdl.handle.net/1911/87774</a>.en_US
dc.identifier.urihttps://hdl.handle.net/1911/87774en_US
dc.language.isoengen_US
dc.rightsCopyright is held by the author, unless otherwise indicated. Permission to reuse, publish, or reproduce the work beyond the bounds of fair use or other exemptions to copyright law must be obtained from the copyright holder.en_US
dc.subjectAlternating direction method of multipliersen_US
dc.subjectConvergence rateen_US
dc.subjectParallel and distributed optimizationen_US
dc.titleGeneralizations of the Alternating Direction Method of Multipliers for Large-Scale and Distributed Optimizationen_US
dc.typeThesisen_US
dc.type.materialTexten_US
thesis.degree.departmentComputational and Applied Mathematicsen_US
thesis.degree.disciplineEngineeringen_US
thesis.degree.grantorRice Universityen_US
thesis.degree.levelDoctoralen_US
thesis.degree.nameDoctor of Philosophyen_US
Files
Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
DENG-DOCUMENT-2014.pdf
Size:
666.42 KB
Format:
Adobe Portable Document Format
License bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
PROQUEST_LICENSE.txt
Size:
5.84 KB
Format:
Plain Text
Description:
No Thumbnail Available
Name:
LICENSE.txt
Size:
2.6 KB
Format:
Plain Text
Description: