Generalized Accelerated Optimization Framework for Big Data Processing

Loading...
Thumbnail Image
Journal Title
Journal ISSN
Volume Title
School of Electrical Engineering | Doctoral thesis (article-based) | Defence date: 2024-09-06
Date
2024
Major/Subject
Mcode
Degree programme
Language
en
Pages
63 + app. 109
Series
Aalto University publication series DOCTORAL THESES, 166/2024
Abstract
Large-scale optimization problems arise in different fields of engineering and science. Due to the large number of parameters and different structures that these problems can have black-box first-order methods are widely used in solving them. Among the existing firstorder methods, the ones that are most widely used are different variants of Fast Gradient Methods (FGM). Such methods are devised in the context of the estimating sequences framework and exhibit desirable properties such as fast convergence rate and low per iteration complexity. In this Thesis, we devise new estimating sequences and show that they can be used to construct accelerated first-order methods. We start by considering the simplest case, i.e., minimizing smooth and convex objective functions. For this class of problems, we present a class of generalized estimating sequences, constructed by exploiting the history of the estimating functions that are obtained during the minimization process. Using these generalized estimating sequences, we devise a new accelerated gradient method and prove that it converges to a tolerated neighborhood of the optimal solution faster than FGM and other first-order methods. We then consider a more general class of optimization problems, namely composite objectives. For this class of problems, we introduce the class of composite estimating sequences, which are obtained by making use of the gradient mapping framework and a tight lower bound on the function that should be minimized. Using these composite estimating sequences, we devise a composite objective accelerated multi-step estimating sequence technique and prove its accelerated convergence rate. Last, embedding the memory term coming from the previous iterates into the composite estimating sequences, we obtain the generalized composite estimating sequences. Using these estimating sequences, we construct another accelerated gradient method and prove its accelerated convergence rate. The methods devised for solving composite objective functions that we introduce in this thesis are also equipped with efficient backtracking line-search strategies, which enable more accurate estimates of the step-size. Our results are validated by a large number of computational experiments on different types of loss functions, wherein both simulated and publicly available real-world datasets are considered. Our numerical experiments also highlight the robustness of our newly introduced methods to the usage of inexact values for of the Lipschitz constant and the strong convexity parameter.
Description
Supervising professor
Vorobyov, Sergiy A., Prof., Aalto University, Department of Information and Communications Engineering, Finland
Thesis advisor
Charalambous, Themistoklis, Prof., University of Cyprus, Cyprus
Keywords
big data, fast gradient methods, FGM
Other note
Parts
  • [Publication 1]: E. Dosti, S. A. Vorobyov, T. Charalambous. Generalizing Nesterov’s Acceleration Framework by Embedding Momentum Into Estimating Sequences: New Algorithm and Bounds. In IEEE International Symposium on Information Theory (ISIT), Helsinki, Finland, 1506-1511, June 2022.
    DOI: 10.1109/ISIT50566.2022.9834684 View at publisher
  • [Publication 2]: E. Dosti, S. A. Vorobyov, T. Charalambous. Embedding a Heavy-Ball type of Momentum into the Estimating Sequences. Journal Submission, March 2024.
  • [Publication 3]: E. Dosti, S. A. Vorobyov, T. Charalambous. A new accelerated gradient-based estimating sequence technique for solving large-scale optimization problems with composite structure. In IEEE 61st Conference on Decision and Control (CDC), Cancun, Mexico, 7516-7521, January 2023.
    DOI: 10.1109/CDC51059.2022.9993313 View at publisher
  • [Publication 4]: E. Dosti, S. A. Vorobyov, T. Charalambous. A new class of composite objective multistep estimating sequence techniques. Signal Processing, 206, 108889, December 2022.
    DOI: 10.1016/j.sigpro.2022.108889 View at publisher
  • [Publication 5]: E. Dosti, S. A. Vorobyov, T. Charalambous. Generalizing the estimating sequences with memory terms for minimizing convex composite functions. Journal Submission, March 2024.
Citation