Upcoming events
Statistical analysis of overparameterized neural networks
For many years, classical learning theory suggested that neural networks with a large num-ber of parameters would overfit their training data and thus generalize poorly to new, unseen data. Contrary to this long-held belief, the empirical success of such networks has been remarkable. However, from a mathematical perspective, the reasons behind their perfor-mance are not fully understood.
In this talk, we consider overparameterized neural networks learned by gradient descent in a statistical setting. We show that an estimator based on an overparameterized neural net-work - trained with a suitable step size and for an appropriate number of gradient descent steps - can be universally consistent. Furthermore, under suitable smoothness assumptions on the regression function, we derive rates of convergence for this estimator.
These results provide new insights into why overparameterized neural networks can
generalize effectively despite their high complexity.
Selina Drews
TU Darmstadt
KIT Zentrum MathSEE
Karlsruher Institut für Technologie
Englerstraße 2
76131 Karlsruhe
Mail: MathSEE ∂ kit edu
https://www.mathsee.kit.edu/
Deep Learning Workshop (Oct 5-6, on-site at TRIANGEL.space)
info and registrationSep 18-20, 2023 at KIT
Info and registrationKCDS Fellows present: Elevator Pitches on PhD projects and previous scientific work + KCDS 1st Birthday Party on June 27, 2023, 13:00-14:00
moreDr. John A. Warwicker (IOR), May 23, 2023, 13:00-14:00h
moreDr. Johannes Bracher (ECON), April 25, 2023, 13:00-14:00h
moreDr. Cihan Ates (ITS), March 28, 2023, 13:00-14:00h
morePD Dr.-Ing. Uwe Ehret (IWG), Feb 28, 2023, 13:00-14:00h
moreKCDS X GRACE Crossover Workshop (Dec 2022, on-site)
more