PulseAugur
LIVE 06:06:24
research · [1 source] ·
0
research

Researchers pinpoint origin of neural network 'Edge of Stability' phenomenon

Researchers have introduced a new concept called the 'edge coupling' to explain the phenomenon known as the Edge of Stability in neural network training. This functional, applied to consecutive iterate pairs, helps to explain why the largest Hessian eigenvalue is driven to the threshold of $2/\eta$ (where $\eta$ is the learning rate) during full-batch gradient descent. The proposed method provides an exact forcing of the Hessian eigenvalue without any gap, offering a more unified explanation for this observed behavior. AI

Summary written by gemini-2.5-flash-lite from 1 source. How we write summaries →

IMPACT Provides a theoretical framework that could lead to more stable and efficient neural network training.

RANK_REASON Academic paper detailing a new theoretical explanation for a phenomenon in neural network training.

Read on arXiv stat.ML →

Researchers pinpoint origin of neural network 'Edge of Stability' phenomenon

COVERAGE [1]

  1. arXiv stat.ML TIER_1 · Elon Litman ·

    The Origin of Edge of Stability

    Full-batch gradient descent on neural networks drives the largest Hessian eigenvalue to the threshold $2/η$, where $η$ is the learning rate. This phenomenon, the Edge of Stability, has resisted a unified explanation: existing accounts establish self-regulation near the edge but d…