Modeling from Features: a Mean-field Framework for Over-parameterized Deep Neural Networks

Cong Fang , Jason Lee , Pengkun Yang , Tong Zhang

[Proceedings link] [PDF]

Session: Neural Networks/Deep Learning (A)

Session Chair: Quanquan Gu

Poster: Poster Session 2

Abstract: This paper proposes a new mean-field framework for over-parameterized deep neural networks (DNNs), which can be used to analyze neural network training. In this framework, a DNN is represented by probability measures and functions over its features (that is, the function values of the hidden units over the training data) in the continuous limit, instead of the neural network parameters as most existing studies have done. This new representation overcomes the degenerate situation where all the hidden units essentially have only one meaningful hidden unit in each middle layer, leading to a simpler representation of DNNs. Moreover, we construct a non-linear dynamics called neural feature flow, which captures the evolution of an over-parameterized DNN trained by Gradient Descent. We illustrate the framework via the Residual Network (Res-Net) architecture. It is shown that when the neural feature flow process converges, it reaches a global minimal solution under suitable conditions.

Summary presentation

Full presentation