SwissMAP Logo
Log in
  • About us
    • Organization
    • Professors
    • Senior Researchers
    • Postdocs
    • PhD Students
    • Alumni
  • News & Events
    • News
    • Events
    • Online Events
    • Videos
    • Newsletters
    • Press Coverage
    • Perspectives Journal
    • Interviews
  • Research
    • Basic Notions
    • Phase III Directions
    • Phases I & II Projects
    • Publications
    • SwissMAP Research Station
  • Awards, Visitors & Vacancies
    • Awards
    • Innovator Prize
    • Visitors
    • Vacancies
  • Outreach & Education
    • Masterclasses & Doctoral Schools
    • Mathscope
    • Maths Club
    • Athena Project
    • ETH Math Youth Academy
    • SPRING
    • Junior Euler Society
    • General Relativity for High School Students
    • Outreach Resources
    • Exhibitions
    • Previous Programs
    • Events in Outreach
    • News in Outreach
  • Equal Opportunities
    • Mentoring Program
    • Financial Support
    • SwissMAP Scholars
    • Events in Equal Opportunities
    • News in Equal Opportunities
  • Contact
    • Corporate Design
  • Basic Notions
  • Phase III Directions
  • Phases I & II Projects
  • Publications
  • SwissMAP Research Station

Freeze and Chaos for DNNs: an NTK view of Batch Normalization, Checkerboard and Boundary Effects

Arthur Jacot, Franck Gabriel, Clément Hongler

11/7/19 Published in : arXiv:1907.05715

In this paper, we analyze a number of architectural features of Deep Neural Networks (DNNs), using the so-called Neural Tangent Kernel (NTK). The NTK describes the training trajectory and generalization of DNNs in the infinite-width limit.
In this limit, we show that for (fully-connected) DNNs, as the depth grows, two regimes appear: "freeze" (also known as "order"), where the (scaled) NTK converges to a constant (slowing convergence), and "chaos", where it converges to a Kronecker delta (limiting generalization).
We show that when using the scaled ReLU as a nonlinearity, we naturally end up in the "freeze". We show that Batch Normalization (BN) avoids the freeze regime by reducing the importance of the constant mode in the NTK. A similar effect is obtained by normalizing the nonlinearity which moves the network to the chaotic regime.
We uncover the same "freeze" and "chaos" modes in Deep Deconvolutional Networks (DC-NNs). The "freeze" regime is characterized by checkerboard patterns in the image space in addition to the constant modes in input space. Finally, we introduce a new NTK-based parametrization to eliminate border artifacts and we propose a layer-dependent learning rate to improve the convergence of DC-NNs.
We illustrate our findings by training DCGANs using our setup. When trained in the "freeze" regime, we see that the generator collapses to a checkerboard mode. We also demonstrate numerically that the generator collapse can be avoided and that good quality samples can be obtained, by tuning the nonlinearity to reach the "chaos" regime (without using batch normalization).

Entire article

Phase I & II research project(s)

  • Statistical Mechanics

Cosmological Number Counts in Einstein and Jordan frames

Equivariant Batalin-Vilkovisky formalism

  • Leading house

  • Co-leading house


The National Centres of Competence in Research (NCCRs) are a funding scheme of the Swiss National Science Foundation

© SwissMAP 2025 - All rights reserved