Stable Nonconvex-Nonconcave Training via Linear Interpolation: Last iterate under cohypomonotonicity

Written by interpolation | Published 2024/03/07
Tech Story Tags: linear-interpolation | nonexpansive-operators | rapp | cohypomonotone-problems | lookahead-algorithms | rapp-and-lookahead | training-gans | nonmonotone-class

TLDRThis paper presents a theoretical analysis of linear interpolation as a principled method for stabilizing (large-scale) neural network training.via the TL;DR App

This paper is available on arxiv under CC 4.0 license.

Authors:

(1) Thomas Pethick, EPFL (LIONS) thomas.pethick@epfl.ch;

(2) Wanyun Xie, EPFL (LIONS) wanyun.xie@epfl.ch;

(3) Volkan Cevher, EPFL (LIONS) volkan.cevher@epfl.ch.

Table of Links

6 Last iterate under cohypomonotonicity

The above lemma allows us to obtain last iterate convergence for IKM on the inexact resolvent by combing the lemma with Theorem C.1.

Remark 6.3. Notice that the rate in Theorem 6.2 has no dependency on ρ. Specifically, it gets rid of the factor γ/(γ + 2ρ) which Gorbunov et al. (2022b, Thm. 3.2) shows is unimprovable for PP. Theorem 6.2 requires that the iterates stays bounded. In Corollary 6.4 we will assume bounded diameter for simplicity, but it is relatively straightforward to show that the iterates can be guaranteed to be bounded by controlling the inexactness (see Lemma E.2).


Written by interpolation | #1 Publication focused exclusively on Interpolation, ie determining value from the existing values in a given data set.
Published by HackerNoon on 2024/03/07