Stochastic Optimization Methods for Modern Machine Learning Problems

Stochastic Optimization Methods for Modern Machine Learning Problems
Author: Yuejiao Sun
Publisher:
Total Pages: 178
Release: 2021
Genre:
ISBN:


Download Stochastic Optimization Methods for Modern Machine Learning Problems Book in PDF, Epub and Kindle

Optimization has been the workhorse of solving machine learning problems. However, the efficiency of these methods remains far from satisfaction to meet the ever-growing demand that arises in modern applications. In this context, the present dissertation will focus on two fundamental classes of machine learning problems: 1) stochastic nested problems, where one subproblem builds upon the solution of others; and, 2) stochastic distributed problems, where the subproblems are coupled through sharing the common variables. One key difficulty of solving stochastic nested problems is that the hierarchically coupled structure makes the computation of (stochastic) gradients, the basic element in first-order optimization machinery, prohibitively expensive or even impossible.We will develop the first stochastic optimization method, which runs in a single-loop manner and achieves the same sample complexity as the stochastic gradient descent method for non-nested problems. One key difficulty of solving stochastic distributed problems is the resource intensity, especially when algorithms are running atresource-limited devices. In this context, we will introduce a class of communication-adaptive stochastic gradient descent (SGD) methods, which adaptively reuse the stale gradients, thus saving communication. We will show that the new algorithms have convergence rates comparable to original SGD and Adam algorithms, but enjoy impressive empirical performance in terms of total communication round reduction.

Optimization for Machine Learning

Optimization for Machine Learning
Author: Suvrit Sra
Publisher: MIT Press
Total Pages: 509
Release: 2012
Genre: Computers
ISBN: 026201646X


Download Optimization for Machine Learning Book in PDF, Epub and Kindle

An up-to-date account of the interplay between optimization and machine learning, accessible to students and researchers in both communities. The interplay between optimization and machine learning is one of the most important developments in modern computational science. Optimization formulations and methods are proving to be vital in designing algorithms to extract essential knowledge from huge volumes of data. Machine learning, however, is not simply a consumer of optimization technology but a rapidly evolving field that is itself generating new optimization ideas. This book captures the state of the art of the interaction between optimization and machine learning in a way that is accessible to researchers in both fields. Optimization approaches have enjoyed prominence in machine learning because of their wide applicability and attractive theoretical properties. The increasing complexity, size, and variety of today's machine learning models call for the reassessment of existing assumptions. This book starts the process of reassessment. It describes the resurgence in novel contexts of established frameworks such as first-order methods, stochastic approximations, convex relaxations, interior-point methods, and proximal methods. It also devotes attention to newer themes such as regularized optimization, robust optimization, gradient and subgradient methods, splitting techniques, and second-order methods. Many of these techniques draw inspiration from other fields, including operations research, theoretical computer science, and subfields of optimization. The book will enrich the ongoing cross-fertilization between the machine learning community and these other fields, and within the broader optimization community.

Stochastic Optimization Methods

Stochastic Optimization Methods
Author: Kurt Marti
Publisher: Springer Nature
Total Pages: 389
Release:
Genre:
ISBN: 3031400593


Download Stochastic Optimization Methods Book in PDF, Epub and Kindle

First-order and Stochastic Optimization Methods for Machine Learning

First-order and Stochastic Optimization Methods for Machine Learning
Author: Guanghui Lan
Publisher: Springer Nature
Total Pages: 591
Release: 2020-05-15
Genre: Mathematics
ISBN: 3030395685


Download First-order and Stochastic Optimization Methods for Machine Learning Book in PDF, Epub and Kindle

This book covers not only foundational materials but also the most recent progresses made during the past few years on the area of machine learning algorithms. In spite of the intensive research and development in this area, there does not exist a systematic treatment to introduce the fundamental concepts and recent progresses on machine learning algorithms, especially on those based on stochastic optimization methods, randomized algorithms, nonconvex optimization, distributed and online learning, and projection free methods. This book will benefit the broad audience in the area of machine learning, artificial intelligence and mathematical programming community by presenting these recent developments in a tutorial style, starting from the basic building blocks to the most carefully designed and complicated algorithms for machine learning.

Reinforcement Learning and Stochastic Optimization

Reinforcement Learning and Stochastic Optimization
Author: Warren B. Powell
Publisher: John Wiley & Sons
Total Pages: 1090
Release: 2022-03-15
Genre: Mathematics
ISBN: 1119815037


Download Reinforcement Learning and Stochastic Optimization Book in PDF, Epub and Kindle

REINFORCEMENT LEARNING AND STOCHASTIC OPTIMIZATION Clearing the jungle of stochastic optimization Sequential decision problems, which consist of “decision, information, decision, information,” are ubiquitous, spanning virtually every human activity ranging from business applications, health (personal and public health, and medical decision making), energy, the sciences, all fields of engineering, finance, and e-commerce. The diversity of applications attracted the attention of at least 15 distinct fields of research, using eight distinct notational systems which produced a vast array of analytical tools. A byproduct is that powerful tools developed in one community may be unknown to other communities. Reinforcement Learning and Stochastic Optimization offers a single canonical framework that can model any sequential decision problem using five core components: state variables, decision variables, exogenous information variables, transition function, and objective function. This book highlights twelve types of uncertainty that might enter any model and pulls together the diverse set of methods for making decisions, known as policies, into four fundamental classes that span every method suggested in the academic literature or used in practice. Reinforcement Learning and Stochastic Optimization is the first book to provide a balanced treatment of the different methods for modeling and solving sequential decision problems, following the style used by most books on machine learning, optimization, and simulation. The presentation is designed for readers with a course in probability and statistics, and an interest in modeling and applications. Linear programming is occasionally used for specific problem classes. The book is designed for readers who are new to the field, as well as those with some background in optimization under uncertainty. Throughout this book, readers will find references to over 100 different applications, spanning pure learning problems, dynamic resource allocation problems, general state-dependent problems, and hybrid learning/resource allocation problems such as those that arose in the COVID pandemic. There are 370 exercises, organized into seven groups, ranging from review questions, modeling, computation, problem solving, theory, programming exercises and a “diary problem” that a reader chooses at the beginning of the book, and which is used as a basis for questions throughout the rest of the book.

Stochastic Optimization Methods

Stochastic Optimization Methods
Author: Kurt Marti
Publisher: Springer Science & Business Media
Total Pages: 332
Release: 2005
Genre: Business & Economics
ISBN: 9783540222729


Download Stochastic Optimization Methods Book in PDF, Epub and Kindle

This text provides a concise overview of stochastic optimization and considers nonlinear optimization problems. Optimization problems arising in practice involve random parameters. For the computation of robust optimal solutions, deterministic substitute problems are needed. Based on the distribution of the random data, and using decision theoretical concepts, optimization problems under stochastic uncertainty are converted into deterministic substitute problems.

Optimization for Machine Learning

Optimization for Machine Learning
Author: Jason Brownlee
Publisher: Machine Learning Mastery
Total Pages: 412
Release: 2021-09-22
Genre: Computers
ISBN:


Download Optimization for Machine Learning Book in PDF, Epub and Kindle

Optimization happens everywhere. Machine learning is one example of such and gradient descent is probably the most famous algorithm for performing optimization. Optimization means to find the best value of some function or model. That can be the maximum or the minimum according to some metric. Using clear explanations, standard Python libraries, and step-by-step tutorial lessons, you will learn how to find the optimum point to numerical functions confidently using modern optimization algorithms.

Understanding and Accelerating the Optimization of Modern Machine Learning

Understanding and Accelerating the Optimization of Modern Machine Learning
Author: Chaoyue Liu (Ph. D. in computer science)
Publisher:
Total Pages: 0
Release: 2021
Genre: Deep learning (Machine learning)
ISBN:


Download Understanding and Accelerating the Optimization of Modern Machine Learning Book in PDF, Epub and Kindle

Over the last decade, we have seen impressive progress of deep learning on a variety of intelligence tasks. The success of deep learning is due, to a great extent, to the remarkable effectiveness of gradient-based optimization methods applied to large neural networks, which is often over-parameterized, i.e., the number of parameters greatly exceeds the number of training samples. However, theoretically, it is still far from clear that why gradient descent algorithms can efficiently optimize the seemingly highly non-convex loss functions (a.k.a. objective functions). In this dissertation, we target to close this gap between the theory and the practice. We first show that certain sufficiently wide neural networks, as typical examples of large non-linear models, exhibit an amazing, while somehow counter-intuitive, phenomenon--transition to linearity. Specifically, these networks can be well approximated by linear models; moreover, they become linear models in the infinite network width limit. Based on this phenomenon, we further provide an optimization theory that describes the loss landscape of over-parameterized machine learning models and further explains the convergence of gradient descent methods on these models. Note that this theory covers both the models that have the "transition to linearity'', and those may not have this property, e.g., wide networks with non-linear output layer. Finally, we prove that, in the stochastic setting, the popularly used Nesterov's momentum does not accelerate the stochastic gradient descent, even for quadratic optimization problems. Furthermore, we propose a new method, MaSS, that provably accelerate SGD in the over-parameterized setting.

Gradient Descent, Stochastic Optimization, and Other Tales

Gradient Descent, Stochastic Optimization, and Other Tales
Author: Jun Lu
Publisher: Eliva Press
Total Pages: 0
Release: 2022-07-22
Genre:
ISBN: 9789994981557


Download Gradient Descent, Stochastic Optimization, and Other Tales Book in PDF, Epub and Kindle

The goal of this book is to debunk and dispel the magic behind the black-box optimizers and stochastic optimizers. It aims to build a solid foundation on how and why the techniques work. This manuscript crystallizes this knowledge by deriving from simple intuitions, the mathematics behind the strategies. This book doesn't shy away from addressing both the formal and informal aspects of gradient descent and stochastic optimization methods. By doing so, it hopes to provide readers with a deeper understanding of these techniques as well as the when, the how and the why of applying these algorithms. Gradient descent is one of the most popular algorithms to perform optimization and by far the most common way to optimize machine learning tasks. Its stochastic version receives attention in recent years, and this is particularly true for optimizing deep neural networks. In deep neural networks, the gradient followed by a single sample or a batch of samples is employed to save computational resources and escape from saddle points. In 1951, Robbins and Monro published A stochastic approximation method, one of the first modern treatments on stochastic optimization that estimates local gradients with a new batch of samples. And now, stochastic optimization has become a core technology in machine learning, largely due to the development of the back propagation algorithm in fitting a neural network. The sole aim of this article is to give a self-contained introduction to concepts and mathematical tools in gradient descent and stochastic optimization.

The Cross-Entropy Method

The Cross-Entropy Method
Author: Reuven Y. Rubinstein
Publisher: Springer Science & Business Media
Total Pages: 316
Release: 2013-03-09
Genre: Computers
ISBN: 1475743211


Download The Cross-Entropy Method Book in PDF, Epub and Kindle

Rubinstein is the pioneer of the well-known score function and cross-entropy methods. Accessible to a broad audience of engineers, computer scientists, mathematicians, statisticians and in general anyone, theorist and practitioner, who is interested in smart simulation, fast optimization, learning algorithms, and image processing.