Time-varying optimization using primal-dual type methods: Solution tracking and model training

dc.contributor.advisor Mingyi . Hong
dc.contributor.advisor Lizhi . Wang
dc.contributor.author Zhang, Yijian
dc.contributor.department Department of Industrial and Manufacturing Systems Engineering
dc.date 2020-02-12T23:02:00.000
dc.date.accessioned 2020-06-30T03:21:07Z
dc.date.available 2020-06-30T03:21:07Z
dc.date.copyright Sun Dec 01 00:00:00 UTC 2019
dc.date.embargo 2001-01-01
dc.date.issued 2019-01-01
dc.description.abstract <p>This dissertation considers general time-varying optimization problems that arise in many network control and machine learning applications. The problem takes the form of minimizing the sum of separable convex/non-convex cost functions subject to equality constraints, and both the objective and constraints are parameterized by some time-varying parameters. To cope with the problem dynamics, we design dynamic/stochastic algorithms based on primal-dual type methods, e.g. alternating direction method of multipliers (ADMM) and primal dual decomposition (PDD). Depending on specific application, our algorithms can accomplish two major tasks: i) continuously track optimal solutions for each time instance; ii) learn the general pattern of given data and produce a unique solution that fits all time-varying parameters.</p> <p>The first part of the dissertation focuses on problems changing optimal solutions. We leverage proximal gradient in the primal steps, and modify the dual steps by adding some perturbation to accomondate the time-varying nature of the problem. We show that, under mild assumptions, the proposed algorithm is able to track the change of problem, meaning it will always stay in a neighborhood around the optimal or approximate optimal solutions for each time instance. Moreover, our analysis indicates an interesting trade-off between solution accuracy and convergence speed. In cases where gradient information is not available or difficult to compute, we develop a suitable time-varying algorithm by only using the function value information (also known as the zeroth-order information). We observe similar performance as gradient based methods and convergence in expectation is proved under suitable assumptions. As an extension of this time-varying framework, static optimization with randomness in updates are discussed with applications in power systems. Specifically, an ADMM-based distributed optimal power flow (OPF) controller for renewable energy source (RES) is designed to steer RES output powers to approximate AC OPF solutions.</p> <p>The second part of the dissertation, we further discover that the time varying framework is also applicable to cases where all changing parameters can fit one unique solution, i.e. training. This type of problem is the core of many machine learning models that aiming at extracting data pattern. We specifically focus on deep neural network (DNN) and model the training of DNN into an equality constrained optimization problem by introducing auxiliary variables for each hidden layer. The resulting formulation is highly nonconvex and time varying in that each time only part of the data is available and as time goes by data comes in sequentially. We design another primal-dual type method called stochastic primal-dual decomposition (PDD) for the neural network training problem. We demonstrate that the developed algorithm is effective by: 1) performing theoretical analysis to show that the stochastic PDD algorithm can compute stationary solution of the training problem; 2) conducting comprehensive comparison with state-of-the-art algorithms and show that the proposed algorithm achieves some early stage advantage, that is, the training error decreases faster in the first few epoches.</p>
dc.format.mimetype application/pdf
dc.identifier archive/lib.dr.iastate.edu/etd/17815/
dc.identifier.articleid 8822
dc.identifier.contextkey 16525317
dc.identifier.s3bucket isulib-bepress-aws-west
dc.identifier.submissionpath etd/17815
dc.identifier.uri https://dr.lib.iastate.edu/handle/20.500.12876/31998
dc.language.iso en
dc.source.bitstream archive/lib.dr.iastate.edu/etd/17815/Zhang_iastate_0097E_18500.pdf|||Fri Jan 14 21:29:26 UTC 2022
dc.subject.disciplines Engineering
dc.subject.disciplines Industrial Engineering
dc.subject.keywords convergence analysis
dc.subject.keywords nonlinear optimization
dc.subject.keywords primal-dual
dc.subject.keywords real-time control
dc.subject.keywords time-varying
dc.subject.keywords tracking and training
dc.title Time-varying optimization using primal-dual type methods: Solution tracking and model training
dc.type dissertation
dc.type.genre dissertation
dspace.entity.type Publication
relation.isOrgUnitOfPublication 51d8b1a0-5b93-4ee8-990a-a0e04d3501b1
thesis.degree.discipline Industrial and Manufacturing Systems Engineering
thesis.degree.level dissertation
thesis.degree.name Doctor of Philosophy
File
Original bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
Zhang_iastate_0097E_18500.pdf
Size:
2.58 MB
Format:
Adobe Portable Document Format
Description: