Researchers have introduced a new concept called the 'edge coupling' to explain the phenomenon known as the Edge of Stability in neural network training. This functional, applied to consecutive iterate pairs, helps to explain why the largest Hessian eigenvalue is driven to the threshold of $2/\eta$ (where $\eta$ is the learning rate) during full-batch gradient descent. The proposed method provides an exact forcing of the Hessian eigenvalue without any gap, offering a more unified explanation for this observed behavior. AI
Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →
IMPACT Provides a theoretical framework that could lead to more stable and efficient neural network training.
RANK_REASON Academic paper detailing a new theoretical explanation for a phenomenon in neural network training.