Princeton University Users: If you would like to view a senior thesis while you are away from campus, you will need to connect to the campus network remotely via the Global Protect virtual private network (VPN). If you are not part of the University requesting a copy of a thesis, please note, all requests are processed manually by staff and will require additional time to process.
 

Publication:

Weak Convergence of L2-Regularized Two-Layer Neural Networks under SGD via Mean Field Theory

No Thumbnail Available

Files

NGUYEN_VINCENT_THESIS.pdf (2.39 MB)

Date

2025-04-28

Journal Title

Journal ISSN

Volume Title

Publisher

Research Projects

Organizational Units

Journal Issue

Abstract

Extending the work of Sirignano and Spiliopoulos (2020), we use mean field theory to study two-layer neural networks with l2-regularization trained under single-sample online stochastic gradient descent. We prove that in the asymptotic regime of both infinite training steps and infinite hidden layer width, such a neural network weakly converges to a deterministic and unique solution that satisfies a partial differential equation of the gradient flow form seen elsewhere in optimal transport and physics. Moreover, we show that the parameters of said neural network, despite being interdependent throughout training, asymptotically become independent. These results are only subject to loose moment bounds at initialization. Our proofs utilize a probabilistic approach on the network's training evolution instead of studying the geometry of the loss surface. We also provide numerical simulation results consistent with our theoretical guarantees.

Description

Keywords

Citation