SwissMAP Logo
Log in
  • About us
    • Organization
    • Professors
    • Senior Researchers
    • Postdocs
    • PhD Students
    • Alumni
  • News & Events
    • News
    • Events
    • Online Events
    • Videos
    • Newsletters
    • Press Coverage
    • Perspectives Journal
    • Interviews
  • Research
    • Basic Notions
    • Phase III Directions
    • Phases I & II Projects
    • Publications
    • SwissMAP Research Station
  • Awards, Visitors & Vacancies
    • Awards
    • Innovator Prize
    • Visitors
    • Vacancies
  • Outreach & Education
    • Masterclasses & Doctoral Schools
    • Mathscope
    • Maths Club
    • Athena Project
    • ETH Math Youth Academy
    • SPRING
    • Junior Euler Society
    • General Relativity for High School Students
    • Outreach Resources
    • Exhibitions
    • Previous Programs
    • Events in Outreach
    • News in Outreach
  • Equal Opportunities
    • Mentoring Program
    • Financial Support
    • SwissMAP Scholars
    • Events in Equal Opportunities
    • News in Equal Opportunities
  • Contact
    • Corporate Design
  • Basic Notions
  • Phase III Directions
  • Phases I & II Projects
  • Publications
  • SwissMAP Research Station

Quantitative convergence of trained single layer neural networks to Gaussian processes

Eloy Mosig, Andrea Agazzi, Dario Trevisan

29/9/25 Published in : arXiv:2509.24544

In this paper, we study the quantitative convergence of shallow neural networks trained via gradient descent to their associated Gaussian processes in the infinite-width limit. While previous work has established qualitative convergence under broad settings, precise, finite-width estimates remain limited, particularly during training. We provide explicit upper bounds on the quadratic Wasserstein distance between the network output and its Gaussian approximation at any training time t \ge 0, demonstrating polynomial decay with network width. Our results quantify how architectural parameters, such as width and input dimension, influence convergence, and how training dynamics affect the approximation error.

Entire article

Phase III direction(s)

  • Statistical Mechanics and Random Structures
  • Differential equations of Mathematical Physics

A multiscale analysis of mean-field transformers in the moderate interaction regime

Global Optimization via Softmin Energy Minimization

  • Leading house

  • Co-leading house


The National Centres of Competence in Research (NCCRs) are a funding scheme of the Swiss National Science Foundation

© SwissMAP 2025 - All rights reserved