ACCELERATED SINGLE-CALL METHODS FOR CON-STRAINED MIN-MAX OPTIMIZATION

Abstract

We study first-order methods for constrained min-max optimization. Existing methods either require two gradient calls or two projections in each iteration, which may be costly in some applications. In this paper, we first show that a variant of the Optimistic Gradient (OG) method, a singlecall single-projection algorithm, has O( 1√ T ) best-iterate convergence rate for inclusion problems with operators that satisfy the weak Minty variation inequality (MVI). Our second result is the first single-call singleprojection algorithm -the Accelerated Reflected Gradient (ARG) method that achieves the optimal O( 1 T ) last-iterate convergence rate for inclusion problems that satisfy negative comonotonicity. Both the weak MVI and negative comonotonicity are well-studied assumptions and capture a rich set of non-convex non-concave min-max optimization problems. Finally, we show that the Reflected Gradient (RG) method, another single-call single-projection algorithm, has O( 1 √ T ) last-iterate convergence rate for constrained convex-concave min-max optimization, answering an open problem of (Hsieh et al., 2019). Our convergence rates hold for standard measures such as the tangent residual and the natural residual.

1. INTRODUCTION

Various Machine Learning applications, from the generative adversarial networks (GANs) (e.g., (Goodfellow et al., 2014; Arjovsky et al., 2017)) , adversarial examples (e.g., (Madry et al., 2017) ), robust optimization (e.g., (Ben-Tal et al., 2009) ), to reinforcement learning (e.g., (Du et al., 2017; Dai et al., 2018) ), can be captured by constrained min-max optimization. Unlike the well-behaved convex-concave setting, these modern ML applications often require solving non-convex non-concave min-max optimization problems in high dimensional spaces. Unfortunately, the general non-convex non-concave setting is intractable even for computing a local solution (Hirsch et al., 1989; Papadimitriou, 1994; Daskalakis et al., 2021) . Motivated by the intractability, researchers turn their attention to non-convex non-concave settings with structure. Significant progress has been made for several interesting structured non-convex non-concave settings, such as the ones that satisfy the weak Minty variation inequality (MVI) (Definition 2) (Diakonikolas et al., 2021; Pethick et al., 2022) and the ones that satisfy the more strict negatively comonotone condition (Definition 3) (Lee & Kim, 2021a; Cai et al., 2022a) . These algorithms are variations of the celebrated extragradient (EG) method (Korpelevich, 1976) , an iterative first-order method. Similar to the extragradient method, these algorithms all require two oracle calls per iteration, which may be costly in practice. We investigate the following important question in this paper: Can we design efficient single-call first-order methods for structured non-convex non-concave min-max optimization? (*) We provide an affirmative answer to the question. We first show that a single-call method known as the Optimistic Gradient (OG) method (Hsieh et al., 2019) is applicable to all non-

