News
Today's seminar will be on ZoomWritten on 16.12.24 (last change on 16.12.24) by Advait Gadhikar Hi everyone,
This is just a reminder that today's seminar will be on Zoom. The link for seminar is: https://cispa-de.zoom.us/j/4747668486?pwd=MVBxMHdMY1REL3h6MDZ0b2k3K0k2dz09
Best Advait |
Office hours before the next seminar.Written on 03.12.24 (last change on 03.12.24) by Advait Gadhikar Hi all, We would like to invite you to an office hour before the next seminar. The aim of this hour would be to discuss potential ideas about the paper for your role with us before you finalize your presentation. Tom (tom.jacobs@cispa.de) and I will be conducting this session next week. Date:… Read more Hi all, We would like to invite you to an office hour before the next seminar. The aim of this hour would be to discuss potential ideas about the paper for your role with us before you finalize your presentation. Tom (tom.jacobs@cispa.de) and I will be conducting this session next week. Date: 12th December, 2024 Time: Thursday 5pm - 6pm Venue: C0 (Room 0.07) In case this time does not work for you, feel free to write us an email and schedule a Zoom chat. Best Advait |
Registration on LSF possibleWritten on 30.10.24 by Rebekka Burkholz Hi everyone, registration on LSF (https://www.lsf.uni-saarland.de) is now feasible with deadline: Nov 18, 2024. This is necessary if you wish you receive credit points for this seminar. Thank you very much!
|
Changed deliverablesWritten on 28.10.24 by Rebekka Burkholz Please note that I have changed the deliverables. As you have more work during the semester with preparing different presentations, I have cut down the seminar paper to 2 pages reflecting on your experience changing roles. |
Link to SlidesWritten on 28.10.24 by Rebekka Burkholz Hello everyone, many thanks for the productive kick-off meeting today. Please find the slides with all required information here: https://docs.google.com/presentation/d/1-OkoglbjTmtzvGUZJAtwByr4wsxagUYL6rZqPH7M9As/edit?usp=sharing ... and do not forget to register on LSF if you wish to earn… Read more Hello everyone, many thanks for the productive kick-off meeting today. Please find the slides with all required information here: https://docs.google.com/presentation/d/1-OkoglbjTmtzvGUZJAtwByr4wsxagUYL6rZqPH7M9As/edit?usp=sharing ... and do not forget to register on LSF if you wish to earn credit points. ;) Looking forward to insightful discussions!
|
Topics out!Written on 28.10.24 (last change on 28.10.24) by Rebekka Burkholz Hi everyone, please indicate your role preferences for the respective papers here: https://docs.google.com/spreadsheets/d/1xKxwPh8GeiWbK7w2SEKQ0tA1e1chmUrsd1p07x2UR-s/edit?gid=0#gid=0
|
Let’s Role Play in the Deep: Do we really need overparameterization in deep learning?
It's time to connect to your inner kid in this fun and engaging seminar format (https://colinraffel.com/blog/role-playing-seminar.md.html): Several students read the same paper, but each student takes on a specific role, which defines the lens through which they contribute to the discussion. Students cycle through roles throughout the course of this seminar on overparameterization in deep learning.
Content: Do we really need overparameterization in deep learning?
Deep learning continues to impress us with breakthroughs across disciplines and is a major driving force behind a multitude of industry innovations like ChatGPT. Most of its successes are achieved by increasingly large neural networks that are trained on massive data sets and still achieve zero training loss. This recent trend to overparameterize neural networks defies classic concepts of statistical learning theory that suggest to avoid overfitting by reducing the number of trainable parameters. We will look into recent explanations of this puzzling phenomenon, discuss related insights, and challenge the modern belief that scaling up neural networks is always the best way to move forward. Are the simplest models always the best choice? And is counting parameters really the best way to measure model complexity? Please join the seminar if you enjoy thinking about this kind of questions.
Important Dates
* Kick-off meeting in the second week of the semester (Oct 28, 2024).
* The presentations will be organized as a bi-weekly meeting during the semester. We usually meet at 17:15 in E9 1 (CISPA main building) in Room 0.07. Participation is mandatory. Please find the schedule below:
Date |
Mentor |
Papers |
Code |
Nov 18 |
Celia (celia.rubio-madrigal@cispa.de) |
||
Dec 2 |
Advait (advait.gadhikar@cispa.de) |
||
Dec 16 (virtual) |
Tom (tom.jacobs@cispa.de) |
||
Jan 6 |
Gowtham (gowtham.abbavaram@cispa.de) |
Rethinking Bias-Variance Trade-off for Generalization of Neural Networks |
|
Jan 20 |
Chao & Dong (chao.zhou@cispa.de, dong.sun@cispa.de) |
On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima |
|
Feb 3 |
Rahul (rahul.nittala@cispa.de) |
Sharpness-aware Minimization for Efficiently Improving Generalization |
https://github.com/google-research/sam / https://github.com/davda54/sam |
Deliverables
*Short presentations from the perspective of your respective role and regular presentations of your findings: (40% of your final grade).
*Discussion: (40%). How actively you engage in discussions during the seminar will also influence your grade.
*Seminar Paper: (20%) You will write a seminar paper (max. 2 pages) on your role play experience and how your perception of papers changed depending on your role