Physics-Inspired Effective Theory of Model Generalization

Abstract
We present GenEFT, an effective theory framework for shedding light on the statics and dynamics of neural network generalization, and illustrate it with graph learning examples. We first investigate the generalization phase transition as data size increases, comparing experimental results with information-theory-based approximations. We then introduce an effective theory for the dynamics of representation learning, where latent-space representations are modeled as interacting particles (βreponsβ), and find that it explains our experimentally observed phase transition between generalization and overfitting as encoder and decoder learning rates are scanned. This highlights the potential of physics-inspired effective theories for bridging the gap between theoretical predictions and practice in machine learning.
Type
Publication
Phys. Rev. E 111, 035307
Click the Cite button above to demo the feature to enable visitors to import publication metadata into their reference management software.
Create your slides in Markdown - click the Slides button to check out the example.
Add the publication’s full text or supplementary notes here. You can use rich formatting such as including code, math, and images.