Zeroth-Order Methods for Convex-Concave Minmax Problems: Applications to Decision-Dependent Risk Minimization
Abstract
Min-max optimization is emerging as a key framework for analyzing problems of robustness to strategically and adversarially generated data. We propose a random reshuffling-based gradient free Optimistic Gradient Descent-Ascent algorithm for solving convex-concave min-max problems with finite sum structure. We prove that the algorithm enjoys the same convergence rate as that of zeroth-order algorithms for convex minimization problems. We further specialize the algorithm to solve distributionally robust, decision-dependent learning problems, where gradient information is not readily available. Through illustrative simulations, we observe that our proposed approach learns models that are simultaneously robust against adversarial distribution shifts and strategic decisions from the data sources, and outperforms existing methods from the strategic classification literature.
Attached Files
Submitted - 2106.09082.pdf
Files
2106.09082.pdf
Files
(838.4 kB)
| Name | Size | Download all |
|---|---|---|
|
md5:a5f6b82e2a5a5a11bab2a636f5061d6f
|
838.4 kB | Preview Download |
Additional details
Identifiers
- Eprint ID
- 110725
- Resolver ID
- CaltechAUTHORS:20210903-213714292
Related works
- Describes
- http://arxiv.org/abs/2106.09082 (URL)
Dates
- Created
-
2021-09-07Created from EPrint's datestamp field
- Updated
-
2023-06-02Created from EPrint's last_modified field