The Generalization Error Bound for A Stochastic Gradient Descent Family bia A Gaussian Approximation Method
Data publikacji: 24 cze 2025
Zakres stron: 251 - 266
Otrzymano: 03 cze 2024
Przyjęty: 16 paź 2024
DOI: https://doi.org/10.61822/amcs-2025-0018
Słowa kluczowe
© 2025 Hao Chen et al., published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 3.0 License.
Recent works have developed model complexity based and algorithm based generalization error bounds to explain how stochastic gradient descent (SGD) methods help over-parameterized models generalize better. However, previous works are limited by their scope of analysis and fail to provide comprehensive explanations. In this paper, we propose a novel Gaussian approximation framework to establish generalization error bounds for the 𝒰-SGD family, which is a class of SGD with asymptotically unbiased and uniformly bounded gradient noise. We study 𝒰-SGD dynamics, and we show both theoretically and numerically that the limiting model parameter distribution tends to be Gaussian, even when the original gradient noise is non-Gaussian. For a 𝒰-SGD family, we establish a desirable iteration number independent generalization error bound at the order of