En poursuivant votre navigation sur ce site, vous acceptez l'utilisation d'un simple cookie d'identification. Aucune autre exploitation n'est faite de ce cookie. OK
1

Gradient descent for wide two-layer neural networks

Bookmarks Report an error
Multi angle
Authors : Bach, Francis (Author of the conference)
CIRM (Publisher )

Loading the player...

Abstract : Neural networks trained to minimize the logistic (a.k.a. cross-entropy) loss with gradient-based methods are observed to perform well in many supervised classification tasks. Towards understanding this phenomenon, we analyze the training and generalization behavior of infinitely wide two-layer neural networks with homogeneous activations. We show that the limits of the gradient flow on exponentially tailed losses can be fully characterized as a max-margin classifier in a certain non-Hilbertian space of functions.

Keywords : optimization; neural networks; machine learning

MSC Codes :
65K05 - Mathematical programming methods
65K10 - Optimization and variational techniques
68T99 - None of the above but in this section
68W99 - None of the above

    Information on the Video

    Film maker : Hennenfent, Guillaume
    Language : English
    Available date : 06/04/2020
    Conference Date : 12/03/2020
    Subseries : Research talks
    arXiv category : Machine Learning ; Optimization and Control ; Statistics Theory
    Mathematical Area(s) : Computer Science ; Control Theory & Optimization ; Probability & Statistics
    Format : MP4 (.mp4) - HD
    Video Time : 00:47:31
    Targeted Audience : Researchers
    Download : https://videos.cirm-math.fr/2020-03-12_Bach.mp4/

Information on the Event

Event Title : Optimization for Machine Learning / Optimisation pour l'apprentissage automatique
Event Organizers : Boyer, Claire ; d'Aspremont, Alexandre ; Gramfort, Alexandre ; Salmon, Joseph ; Villar, Soledad
Dates : 09/03/2020 - 13/03/2020
Event Year : 2020
Event URL : https://conferences.cirm-math.fr/2133.html

Citation Data

DOI : 10.24350/CIRM.V.19622703
Cite this video as: Bach, Francis (2020). Gradient descent for wide two-layer neural networks. CIRM. Audiovisual resource. doi:10.24350/CIRM.V.19622703
URI : http://dx.doi.org/10.24350/CIRM.V.19622703

See Also

Bibliography



Imagette Video

Bookmarks Report an error