Curvature-Adaptive Perturbation and Subspace Descent for Robust Saddle Point Escape in High-Dimensional Optimization

Ronald Katende, Henry Kasumba

公開日: 2024/9/19

Abstract

High-dimensional non-convex optimization problems in engineering design, control, and learning are often hindered by saddle points, flat plateaus, and strongly anisotropic curvature. This paper develops a unified, curvature-adaptive framework that combines stochastic perturbations, adaptive learning rates, and randomized subspace descent to enhance escape efficiency and scalability. We show theoretically that gradient flow almost surely avoids strict saddles, with escape probability increasing exponentially in dimension. For noise-perturbed gradient descent, we derive explicit escape-time bounds that depend on local curvature and noise magnitude. Adaptive step sizes further reduce escape times by responding to local gradient variability, while randomized subspace descent preserves descent directions in low-dimensional projections and ensures global convergence with logarithmic dependence on dimension. Numerical experiments on nonlinear and constrained benchmarks validate these results, demonstrating faster escape, improved robustness to ill-conditioning, and lower total runtime compared to standard first- and second-order methods. The proposed approach offers practical tools for large-scale engineering optimization tasks where curvature, noise, and dimensionality interplay critically.

Curvature-Adaptive Perturbation and Subspace Descent for Robust Saddle Point Escape in High-Dimensional Optimization | SummarXiv | SummarXiv