Dear all,

 This Friday we have Tomer Koren from Tel Aviv University in the thematic seminar:

Tomer Koren (Tel Aviv University, https://tomerkoren.github.io/)

Friday March 11, 16h00-17h00
Online on Zoom: https://uva-live.zoom.us/j/85690850169
Meeting ID: 856 9085 0169

Please also join for online drinks after the talk.

Benign Underfitting of Stochastic Gradient Descent

We study to what extent may stochastic gradient descent (SGD) be understood as a ``conventional'' learning rule that achieves generalization performance by obtaining a good fit to training data.  We consider the fundamental stochastic convex optimization framework, where SGD is classically known to minimize the population risk at an optimal rate, and prove that, surprisingly, there exist problem instances where the SGD solution exhibits both empirical risk and generalization gap lower bounded by a universal constant.  Consequently, it turns out that SGD is not algorithmically stable in any sense, and its generalization ability cannot be explained by uniform convergence or any other currently known generalization bound technique for that matter (other than that of its classical analysis).  Time permitting, we will discuss related results for with-replacement SGD, multi-epoch SGD, and full-batch gradient descent.

Based on joint works with Idan Amir, Roi Livni, Yishay Mansour and Uri Sherman.


Seminar organizers:
Tim van Erven
Botond Szabo

https://mschauer.github.io/StructuresSeminar/

Upcoming talks:

Mar. 25, Nicolò Cesa-Bianchi, Università degli Studi di Milano
Apr. 22, Tor Lattimore, DeepMind
Jun. 10, Julia Olkhovskaya, Vrije Universiteit
-- 
Tim van Erven <tim@timvanerven.nl>
www.timvanerven.nl