Javascript must be enabled to continue!
Nonconvex min-max optimization in deep learning
View through CrossRef
Nonconvex min-max optimization receives increasing attention in modern machine learning, especially in the context of deep learning. Examples include stochastic AUC maximization with deep neural networks and Generative Adversarial Nets (GANs), which correspond to a nonconvex-concave and nonconvex-nonconcave min-max problem respectively. The classical theory of min-max optimization mainly focuses on convex-concave setting, which is not applicable for deep learning applications with nonconvex min-max formulation. A natural question is proposed---how to design provably efficient algorithms for nonconvex min-max problems in deep learning? To answer this question, this dissertation focuses on the following four concrete aspects:
First, we consider the problem of stochastic AUC maximization problem with deep neural networks as preditive models. Building on the saddle point reformulation of a surrogated loss of AUC, the problem can be cast into a {\it non-convex concave} min-max problem. We explore Polyak-Lojasiewicz (PL) condition that has been proved and observed in deep learning, and develop new stochastic algorithms with even faster convergence rate and more practical step size scheme.
Second, we consider the first-order convergence theory for weakly-convex-weakly-concave min-max problems. We propose an algorithmic framework motivated by the inexact proximal point method, where the weakly monotone variational inequality (VI) corresponding to the original min-max problem is solved through approximately solving a sequence of strongly monotone VIs constructed by adding a strongly monotone mapping to the original gradient mapping. We prove first-order non-asymptotic convergence to a nearly stationary point of the original min-max problem in polynomial time.
Third, we consider a class of nonconvex-nonconcave min-max problems with the focus on GAN training. Although adaptive gradient methods with alternate update empirically work well in training GANs, it requires expensive tuning efforts, lacks theoretical convergence guarantees and might diverge in practice. To bridge the gap, we design an adaptive gradient algorithm for training GANs with provably faster convergence than its non-adaptive counterpart.
Fourth, we aim to consider large-scale GAN training in a decentralized distributed manner. Decentralized parallel algorithms are robust to network bandwidth and latency compared with its centralized counterpart and it has merits for protecting users' privacy, but decentralized algorithms for nonconvex-nonconcave min-max optimization are not considered in the existing literature. We propose and analyze a decentralized algorithm for train GANs, and show its provable convergence to first-order stationary point in polynomial time.
The University of Iowa
Title: Nonconvex min-max optimization in deep learning
Description:
Nonconvex min-max optimization receives increasing attention in modern machine learning, especially in the context of deep learning.
Examples include stochastic AUC maximization with deep neural networks and Generative Adversarial Nets (GANs), which correspond to a nonconvex-concave and nonconvex-nonconcave min-max problem respectively.
The classical theory of min-max optimization mainly focuses on convex-concave setting, which is not applicable for deep learning applications with nonconvex min-max formulation.
A natural question is proposed---how to design provably efficient algorithms for nonconvex min-max problems in deep learning? To answer this question, this dissertation focuses on the following four concrete aspects:
First, we consider the problem of stochastic AUC maximization problem with deep neural networks as preditive models.
Building on the saddle point reformulation of a surrogated loss of AUC, the problem can be cast into a {\it non-convex concave} min-max problem.
We explore Polyak-Lojasiewicz (PL) condition that has been proved and observed in deep learning, and develop new stochastic algorithms with even faster convergence rate and more practical step size scheme.
Second, we consider the first-order convergence theory for weakly-convex-weakly-concave min-max problems.
We propose an algorithmic framework motivated by the inexact proximal point method, where the weakly monotone variational inequality (VI) corresponding to the original min-max problem is solved through approximately solving a sequence of strongly monotone VIs constructed by adding a strongly monotone mapping to the original gradient mapping.
We prove first-order non-asymptotic convergence to a nearly stationary point of the original min-max problem in polynomial time.
Third, we consider a class of nonconvex-nonconcave min-max problems with the focus on GAN training.
Although adaptive gradient methods with alternate update empirically work well in training GANs, it requires expensive tuning efforts, lacks theoretical convergence guarantees and might diverge in practice.
To bridge the gap, we design an adaptive gradient algorithm for training GANs with provably faster convergence than its non-adaptive counterpart.
Fourth, we aim to consider large-scale GAN training in a decentralized distributed manner.
Decentralized parallel algorithms are robust to network bandwidth and latency compared with its centralized counterpart and it has merits for protecting users' privacy, but decentralized algorithms for nonconvex-nonconcave min-max optimization are not considered in the existing literature.
We propose and analyze a decentralized algorithm for train GANs, and show its provable convergence to first-order stationary point in polynomial time.
Related Results
[RETRACTED] Keto Max Power - BURN FATINSTEAD OF CARBS with Keto Max Power! v1
[RETRACTED] Keto Max Power - BURN FATINSTEAD OF CARBS with Keto Max Power! v1
[RETRACTED]Keto Max Power Reviews: Warning! Don’t Buy Dragons Den Pills Fast Until You Read This UK Latest Report Weight gain’s principle of “energy intake exceeding energy spent”...
CONTINUOUS COMPRESSION WITHOUT DEFIBRILLATION FAVOURED NO SHORT-TERM SURVIVAL IN PROLONGED VENTRICULAR FIBRILLATION
CONTINUOUS COMPRESSION WITHOUT DEFIBRILLATION FAVOURED NO SHORT-TERM SURVIVAL IN PROLONGED VENTRICULAR FIBRILLATION
Objectives
Aims: During the 2005 American Heart Association (AHA) Consensus Conference, compression first versus defibrillation first for sudden cardiac arrest wi...
[RETRACTED] Optimal Max Keto - Does It ReallyWork? v1
[RETRACTED] Optimal Max Keto - Does It ReallyWork? v1
[RETRACTED]Shedding the unwanted weight and controlling the calories of your body is the most challenging and complicated process. As we start aging, we have to deal with lots of...
Inadekvát, aránytalan sinuscsomó-tachycardia: egy régi szívritmuszavar új megvilágításban (II.)
Inadekvát, aránytalan sinuscsomó-tachycardia: egy régi szívritmuszavar új megvilágításban (II.)
Összefoglaló.
Bevezetés: Az inadekvát, aránytalan sinuscsomó-tachycardia a
szív nomotop ingerképzési zavarával járó, nem ritka klinikai szin...
Deep convolutional neural network and IoT technology for healthcare
Deep convolutional neural network and IoT technology for healthcare
Background Deep Learning is an AI technology that trains computers to analyze data in an approach similar to the human brain. Deep learning algorithms can find complex patterns in ...
Effect of Aminophylline on the Protective Action of Common Antiepileptic Drugs Against Electroconvulsions in Mice
Effect of Aminophylline on the Protective Action of Common Antiepileptic Drugs Against Electroconvulsions in Mice
Summary: The increasing amount of data tends to suggest that adenosine‐mediated inhibition may play a role in the anticonvulsant activity of a number of antiepileptic drugs. Conse...
Initial Experience with Pediatrics Online Learning for Nonclinical Medical Students During the COVID-19 Pandemic
Initial Experience with Pediatrics Online Learning for Nonclinical Medical Students During the COVID-19 Pandemic
Abstract
Background: To minimize the risk of infection during the COVID-19 pandemic, the learning mode of universities in China has been adjusted, and the online learning o...
A new type bionic global optimization: Construction and application of modified fruit fly optimization algorithm
A new type bionic global optimization: Construction and application of modified fruit fly optimization algorithm
Fruit fly optimization algorithm, which is put forward through research on the act of foraging and observing groups of fruit flies, has some merits such as simplified operation, st...

