PulseAugur
LIVE 09:02:12
research · [3 sources] ·
0
research

New research refines SGD generalization bounds and covariance estimation

Researchers have developed new methods to analyze the generalization capabilities of Stochastic Gradient Descent (SGD) in machine learning. One paper introduces predictable history-adaptive virtual perturbations, allowing for more accurate generalization bounds by accounting for adaptive noise geometries that depend on the optimization history. Another study examines the high-dimensional scaling limits of online SGD in single-layer networks, revealing how critical step sizes and information exponents influence sample complexity and the emergence of stochastic fluctuations. AI

Summary written by gemini-2.5-flash-lite from 3 sources. How we write summaries →

IMPACT These theoretical advancements in understanding SGD could lead to more robust and efficient training methods for future machine learning models.

RANK_REASON The cluster contains two academic papers on theoretical aspects of machine learning algorithms.

Read on arXiv stat.ML →

COVERAGE [3]

  1. arXiv cs.LG TIER_1 · Mohammad Partohaghighi ·

    Information-Theoretic Generalization Bounds for Stochastic Gradient Descent with Predictable Virtual Noise

    arXiv:2605.00064v1 Announce Type: new Abstract: Information-theoretic generalization bounds analyze stochastic optimization by relating expected generalization error to the mutual information between learned parameters and training data. Virtual perturbation analyses of SGD add a…

  2. arXiv stat.ML TIER_1 · Parsa Rangriz ·

    Limit Theorems for Stochastic Gradient Descent in High-Dimensional Single-Layer Networks

    arXiv:2511.02258v2 Announce Type: replace Abstract: This paper studies the high-dimensional scaling limits of online stochastic gradient descent (SGD). Building on the recent work of Ben Arous, Gheissari, and Jagannath on the effective dynamics of SGD, we study the critical scali…

  3. arXiv stat.ML TIER_1 · Wei Biao Wu ·

    Refining Covariance Matrix Estimation in Stochastic Gradient Descent Through Bias Reduction

    We study online inference and asymptotic covariance estimation for the stochastic gradient descent (SGD) algorithm. While classical methods (such as plug-in and batch-means estimators) are available, they either require inaccessible second-order (Hessian) information or suffer fr…