Platform

  • Search
  • Seminars
  • Conferences
  • Jobs

Resources

  • Submit Content
  • About Us

© 2025 World Wide

Open knowledge for all • Started with World Wide Neuro • A 501(c)(3) Non-Profit Organization

Analytics consent required

World Wide relies on analytics signals to operate securely and keep research services available. Accept to continue, or leave the site.

Review the Privacy Policy for details about analytics processing.

World Wide
SeminarsConferencesWorkshopsCoursesJobsMapsFeedLibrary
← Back

Implicit Bias Sgd Deep

Back to SeminarsBack
Seminar✓ Recording AvailableNeuroscience

On the implicit bias of SGD in deep learning

Amir Globerson

Tel Aviv University

Schedule
Wednesday, October 20, 2021

Showing your local timezone

Schedule

Wednesday, October 20, 2021

1:00 AM America/New_York

Watch recording
Host: van Vreeswijk TNS

Seminar location

Seminar location

Not provided

No geocoded details are available for this content yet.

Watch the seminar

Recording provided by the organiser.

Event Information

Format

Recorded Seminar

Recording

Available

Host

van Vreeswijk TNS

Duration

70.00 minutes

Seminar location

Seminar location

Not provided

No geocoded details are available for this content yet.

World Wide map

Abstract

Tali's work emphasized the tradeoff between compression and information preservation. In this talk I will explore this theme in the context of deep learning. Artificial neural networks have recently revolutionized the field of machine learning. However, we still do not have sufficient theoretical understanding of how such models can be successfully learned. Two specific questions in this context are: how can neural nets be learned despite the non-convexity of the learning problem, and how can they generalize well despite often having more parameters than training data. I will describe our recent work showing that gradient-descent optimization indeed leads to 'simpler' models, where simplicity is captured by lower weight norm and in some cases clustering of weight vectors. We demonstrate this for several teacher and student architectures, including learning linear teachers with ReLU networks, learning boolean functions and learning convolutional pattern detection architectures.

Topics

ReLU networksartificial neural networksconvolutional architecturesdeep learninggeneralizationimplicit biasnon-convexitystochastic gradient descentweight norm

About the Speaker

Amir Globerson

Tel Aviv University

Contact & Resources

Personal Website

scholar.google.com/citations

@amirgloberson

Follow on Twitter/X

twitter.com/amirgloberson

Related Seminars

Seminar64% match - Relevant

Rethinking Attention: Dynamic Prioritization

neuro

Decades of research on understanding the mechanisms of attentional selection have focused on identifying the units (representations) on which attention operates in order to guide prioritized sensory p

Jan 6, 2025
George Washington University
Seminar64% match - Relevant

The Cognitive Roots of the Problem of Free Will

neuro

Jan 7, 2025
Bielefeld & Amsterdam
Seminar64% match - Relevant

The neural basis of exploration and decision-making in individuals and groups

neuro

Jan 8, 2025
Max Planck Institute of Animal Behaviour, Konstanz
World Wide calendar

World Wide highlights

December 2025 • Syncing the latest schedule.

View full calendar
Awaiting featured picks
Month at a glance

Upcoming highlights