The Generalization Error Bound for A Stochastic Gradient Descent Family bia A Gaussian Approximation Method
Publié en ligne: 24 juin 2025
Pages: 251 - 266
Reçu: 03 juin 2024
Accepté: 16 oct. 2024
DOI: https://doi.org/10.61822/amcs-2025-0018
Mots clés
© 2025 Hao Chen et al., published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 3.0 License.
Recent works have developed model complexity based and algorithm based generalization error bounds to explain how stochastic gradient descent (SGD) methods help over-parameterized models generalize better. However, previous works are limited by their scope of analysis and fail to provide comprehensive explanations. In this paper, we propose a novel Gaussian approximation framework to establish generalization error bounds for the đ°-SGD family, which is a class of SGD with asymptotically unbiased and uniformly bounded gradient noise. We study đ°-SGD dynamics, and we show both theoretically and numerically that the limiting model parameter distribution tends to be Gaussian, even when the original gradient noise is non-Gaussian. For a đ°-SGD family, we establish a desirable iteration number independent generalization error bound at the order of