A Fine-Grained Spectral Perspective on Neural Networks

  • Greg Yang ,
  • Hadi Salman

Eighth International Conference on Learning Representations (ICLR 2020) |

Submitted to ICLR 2020

PDF

Are neural networks biased toward simple functions? Does depth always help learn more complex features? Is training the last layer of a network as good as training all layers? These questions seem unrelated at face value, but in this work we give all of them a common treatment from the spectral perspective. We will study the spectra of the *Conjugate Kernel*, CK, (also called the *Neural Network-Gaussian Process Kernel*), and the *Neural Tangent Kernel*, NTK. Roughly, the CK and the NTK tell us respectively “what a network looks like at initialization”and “what a network looks like during and after training.” Their spectra then encode valuable information about the initial distribution and the training and generalization properties of neural networks. By analyzing the eigenvalues, we lend novel insights into the questions put forth at the beginning, and we verify these insights by extensive experiments of neural networks. We believe the computational tools we develop here for analyzing the spectra of CK and NTK serve as a solid foundation for future studies of deep neural networks. The open-sourced code for generating the plots is available for download from Github below.

Related Tools

NNspectra

July 26, 2019

This repo is a companion to the paper "A Fine-Grained Spectral Perspective on Neural Networks".