On the Stability and Generalization of First-order Bilevel Minimax Optimization
View PDF HTML (experimental) Abstract:Bilevel optimization and bilevel minimax optimization have recently emerged as unifying frameworks for a range of machine-learning tasks, including hyperparameter optimization and reinforcement learning. The existing literature focuses on empirical efficiency and convergence guarantees, leaving a critical theoretical gap in understanding how well these algorithms generalize. To bridge this gap, we provide the first systematic generalization analysis for first-order gradient-based bilevel minimax solvers with lower-level minimax problems. Specifically, by leveraging algorithmic stability arguments, we derive fine-grained generalization bounds for three representative algorithms, including single-timescale stochastic gradient descent-ascent, and two variants of two-timescale stochastic gradient descent-ascent. Our results reveal a precise trade-off among algorithmic stability, generalization gaps, and practical settings. Furthermore, extensive empirical evaluations corroborate our theoretical insights on realistic optimization tasks with bilevel minimax structures. Subjects: Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Machine Learning (stat.ML) Cite as: arXiv:2604.20115 [cs.LG] (or arXiv:2604.20115v1 [cs.LG] for this version) https://doi.org/10.48550/arXiv.2604.20115 arXiv-issued DOI via DataCite (pending registration) Submission history From: Xuelin Zhang [view email] [v1] Wed, 22 Apr 2026 02:27:24 UTC (152 KB)
No replies yet. Be first.