Large-scale Optimization Methods for Data-science Applications

Large-scale Optimization Methods for Data-science Applications
Author :
Publisher :
Total Pages : 211
Release :
ISBN-10 : OCLC:1117775104
ISBN-13 :
Rating : 4/5 ( Downloads)

Book Synopsis Large-scale Optimization Methods for Data-science Applications by : Haihao Lu (Ph.D.)

Download or read book Large-scale Optimization Methods for Data-science Applications written by Haihao Lu (Ph.D.) and published by . This book was released on 2019 with total page 211 pages. Available in PDF, EPUB and Kindle. Book excerpt: In this thesis, we present several contributions of large scale optimization methods with the applications in data science and machine learning. In the first part, we present new computational methods and associated computational guarantees for solving convex optimization problems using first-order methods. We consider general convex optimization problem, where we presume knowledge of a strict lower bound (like what happened in empirical risk minimization in machine learning). We introduce a new functional measure called the growth constant for the convex objective function, that measures how quickly the level sets grow relative to the function value, and that plays a fundamental role in the complexity analysis. Based on such measure, we present new computational guarantees for both smooth and non-smooth convex optimization, that can improve existing computational guarantees in several ways, most notably when the initial iterate is far from the optimal solution set. The usual approach to developing and analyzing first-order methods for convex optimization always assumes that either the gradient of the objective function is uniformly continuous (in the smooth setting) or the objective function itself is uniformly continuous. However, in many settings, especially in machine learning applications, the convex function is neither of them. For example, the Poisson Linear Inverse Model, the D-optimal design problem, the Support Vector Machine problem, etc. In the second part, we develop a notion of relative smoothness, relative continuity and relative strong convexity that is determined relative to a user-specified "reference function" (that should be computationally tractable for algorithms), and we show that many differentiable convex functions are relatively smooth or relatively continuous with respect to a correspondingly fairly-simple reference function. We extend the mirror descent algorithm to our new setting, with associated computational guarantees. Gradient Boosting Machine (GBM) introduced by Friedman is an extremely powerful supervised learning algorithm that is widely used in practice -- it routinely features as a leading algorithm in machine learning competitions such as Kaggle and the KDDCup. In the third part, we propose the Randomized Gradient Boosting Machine (RGBM) and the Accelerated Gradient Boosting Machine (AGBM). RGBM leads to significant computational gains compared to GBM, by using a randomization scheme to reduce the search in the space of weak-learners. AGBM incorporate Nesterov's acceleration techniques into the design of GBM, and this is the first GBM type of algorithm with theoretically-justified accelerated convergence rate. We demonstrate the effectiveness of RGBM and AGBM over GBM in obtaining a model with good training and/or testing data fidelity..

Large-scale Optimization Methods for Data-science Applications Related Books

Large-scale Optimization Methods for Data-science Applications
Language: en
Pages: 211
Authors: Haihao Lu (Ph.D.)
Categories:
Type: BOOK - Published: 2019 - Publisher:

GET EBOOK

In this thesis, we present several contributions of large scale optimization methods with the applications in data science and machine learning. In the first pa
Large-scale Optimization Methods
Language: en
Pages: 0
Authors: Nuri Denizcan Vanli
Categories:
Type: BOOK - Published: 2021 - Publisher:

GET EBOOK

Large-scale optimization problems appear quite frequently in data science and machine learning applications. In this thesis, we show the efficiency of coordinat
Optimization for Data Analysis
Language: en
Pages: 239
Authors: Stephen J. Wright
Categories: Computers
Type: BOOK - Published: 2022-04-21 - Publisher: Cambridge University Press

GET EBOOK

A concise text that presents and analyzes the fundamental techniques and methods in optimization that are useful in data science.
Optimization for Machine Learning
Language: en
Pages: 509
Authors: Suvrit Sra
Categories: Computers
Type: BOOK - Published: 2012 - Publisher: MIT Press

GET EBOOK

An up-to-date account of the interplay between optimization and machine learning, accessible to students and researchers in both communities. The interplay betw
Big Data Optimization: Recent Developments and Challenges
Language: en
Pages: 492
Authors: Ali Emrouznejad
Categories: Technology & Engineering
Type: BOOK - Published: 2016-05-26 - Publisher: Springer

GET EBOOK

The main objective of this book is to provide the necessary background to work with big data by introducing some novel optimization algorithms and codes capable