News
Exam gradesWritten on 02.04.24 by Sebastian Stich Dear students, We entered the grades into LSF. All of you passed, congrats! Exam review If you want to review your exam, please schedule a meeting with us before April 19 (we can also send you a scanned copy of your exam if you prefer that). Details on grading It was possible to get 60… Read more Dear students, We entered the grades into LSF. All of you passed, congrats! Exam review If you want to review your exam, please schedule a meeting with us before April 19 (we can also send you a scanned copy of your exam if you prefer that). Details on grading It was possible to get 60 points in the exam. We converted the exam points to the 0-10 scale we used for the projects by dividing your exam points by 5.5. Course wrap-up Thank you for participating in this course. We recognize that some materials may have been provided later than expected. Your feedback is crucial and will guide our improvements to the course materials (we have factored this into the grade scale and set a generous threshold for passing). We hope you gained valuable knowledge from this course, and we appreciate your understanding. We look forward to your presence again in subsequent lectures, seminars, or possibly in a role as a research assistant (Hiwi) in our labs. |
Consultation on Wednesday, March 13thWritten on 12.03.24 by Tatjana Chavdarova Dear students, |
Exam Announcement: March 15, 2024Written on 06.03.24 by Tatjana Chavdarova Dear students, Dear students, |
Second project pointsWritten on 28.02.24 by Tatjana Chavdarova Dear students, The scores for your second project have been uploaded to the CMS, scaled up to 10 as with the first project. |
Consultation today (Feb 15) 14-15hWritten on 15.02.24 by Tatjana Chavdarova Dear all,
|
GML Guest lecture Feb 6th: The Complexity of Constrained Min-Max OptimizationWritten on 30.01.24 by Tatjana Chavdarova Dear all, Dear all, Abstract: Despite its important applications in Machine Learning, min-max optimization of objective functions that are nonconvex-nonconcave remains elusive. We show that an approximate local min-max point of large enough approximation is guaranteed to exist, but finding one such point is PPAD-complete. The same is true of computing an approximate fixed point of the (Projected) Gradient Descent/Ascent update dynamics. An important byproduct of our proof is to establish an unconditional hardness result in the Nemirovsky-Yudin oracle optimization model. In particular, we show that, given oracle access to some function and its gradient, every algorithm that finds an approximate local min-max point needs to make a number of queries that is exponential in the dimension or the accuracy. This comes in sharp contrast to minimization problems, where finding approximate local minima in the same setting can be done with Projected Gradient Descent polynomially many queries. Our result is the first to show an exponential separation between these two fundamental optimization problems in the oracle model. Joint work with Constantinos Daskalakis and Stratis Skoulakis. Short-Bio: Manolis Zampetakis is currently an Assistant Professor at the Computer Science Department of Yale University. Before that, he was a post-doctoral researcher at the EECS Department of UC Berkeley working with Michael Jordan. He received his PhD from the EECS Department at MIT where he was advised by Constantinos Daskalakis. He has been awarded the Google PhD Fellowship and the ACM SIGEcom Doctoral Dissertation Award. He works on the foundations of machine learning (ML), statistics, and data science, with a focus on statistical analysis from biased data, optimization methods in multi-agent environments, and convergence properties of popular heuristic methods. |
Presentation ScheduleWritten on 25.01.24 (last change on 25.01.24) by Sebastian Stich Dear all, We have added the tentative schedule for the presentations next Tuesday to the project instruction document (see materials).
|
Projects UpdateWritten on 22.01.24 by Sebastian Stich Dear all, Project 2: This is a kind reminder to submit a draft of Project 2 by end of today. Please note that the draft will not be graded. However, at the very least, please submit a PDF that includes the title and your names. The more polished the draft is, the better. This will help us to… Read more Dear all, Project 2: This is a kind reminder to submit a draft of Project 2 by end of today. Please note that the draft will not be graded. However, at the very least, please submit a PDF that includes the title and your names. The more polished the draft is, the better. This will help us to schedule the final presentations. Project 1: We would like to apologize for the delay in grading Project 1. We can confirm that all of the properly submitted projects have achieved a passing grade, but we will need a few more days to finalize the grades. |
Tomorrow's Lecture Rescheduled to Thu, Jan 18Written on 15.01.24 by Tatjana Chavdarova Dear students, Dear students, Looking forward to seeing you on Thursday, |
Second project instructions postedWritten on 21.12.23 by Tatjana Chavdarova Dear all, Instructions for the second project and a file containing example topics are available in the materials tab.
|
Written on 20.12.23 by Tatjana Chavdarova Dear students, Several of you have reached out that you couldn't attend the lectures due to scheduling conflicts with other courses. In response, we will offer either additional sessions, recordings, or lecture summaries between the end of the courses and the exam. Feel free to email us with… Read more Dear students, Several of you have reached out that you couldn't attend the lectures due to scheduling conflicts with other courses. In response, we will offer either additional sessions, recordings, or lecture summaries between the end of the courses and the exam. Feel free to email us with specific requests regarding which lectures or material you would like us to review. It's important to note, however, that the deadline for completing the first project remains unchanged. We have extended the first project application deadline to 23:59 on December 23rd, 2023, and the submission deadline is 23:59 on December 28th, 2023. Best regards, |
Exercise session Thu 21 Dec 13h30Written on 20.12.23 by Tatjana Chavdarova Dear all, You will find the Zoom link under the materials tab. Dear all, You will find the Zoom link under the materials tab. |
Exam date:Written on 11.12.23 by Tatjana Chavdarova Dear all, |
GML First ProjectWritten on 05.12.23 by Tatjana Chavdarova Dear all, If you've already applied for the first project:
Dear all, If you've already applied for the first project:
If you haven't applied yet but plan to do so:
Wishing you a pleasant rest of the week, |
Scheduling exam dateWritten on 04.12.23 by Tatjana Chavdarova Dear all, |
No lecture this weekWritten on 04.12.23 by Tatjana Chavdarova Dear all, Tomorrow, on December 5th, there will be no lecture. Unfortunately, due to unforeseen circumstances, we have to cancel it for this week only. We will resume next week with both the lecture and an exercise session immediately following it. Best, |
Registration for the first projectWritten on 27.11.23 by Tatjana Chavdarova Dear all, |
Zoom exercise session - Thursday Nov 23, 13.30hWritten on 20.11.23 by Sebastian Stich Dear students, We are offering the first online exercise session/office hour this week on Thursday, 13.30h (the link will be added under 'materials'). You are cordially invited to ask any questions you may have, especially about the first part (optimization), but of course also about the more… Read more Dear students, We are offering the first online exercise session/office hour this week on Thursday, 13.30h (the link will be added under 'materials'). You are cordially invited to ask any questions you may have, especially about the first part (optimization), but of course also about the more recent content. (Note that the structure or content will not be different from the 'live' session offered on Tuesdays. It is just an opportunity for people who cannot attend on Tuesdays to ask questions). The zoom link for tomorrow's lecture is again "Zoom link (tatjana)." |
Lecture on Tuesday, Nov. 14th, 2023Written on 14.11.23 by Tatjana Chavdarova Dear all, |
First exercise sessionWritten on 06.11.23 by Sebastian Stich Dear students, Tomorrow we will have our first exercise session. We will discuss questions you might have on the optimization part (a second session will be offered on zoom next week). The zoom link to attend the lecture is the same as last week. |
First lecture on Tuesday, Oct. 24th, 2023Written on 23.10.23 (last change on 24.10.23) by Tatjana Chavdarova Dear students, Place: in CISPA lecture hall. E9.1 (CISPA building), room 0.05 (main lecture room) Dear students, Place: in CISPA lecture hall. E9.1 (CISPA building), room 0.05 (main lecture room)
|
Games in Machine Learning
A typical deep learning (DL) pipeline involves defining a utility or loss function and training a set of parameters to optimize it. This approach has limitations. The specified objective may fail to capture the desired behavior of the model, resulting in high-performing models that struggle to make accurate predictions when faced with slightly modified inputs or that learn irrelevant correlations, ultimately leading to subpar performance. This is solved through a technique called robustification, which employs a min-max objective. In this approach, two models, or players, are trained jointly. Each player has a different real-valued objective that depends on the parameters of both players. Consequently, in the field of machine learning, it becomes necessary to understand how to solve two-player games using iterative, gradient-based methods to find the Nash Equilibrium solution. Furthermore, many problems in machine learning inherently involve multiple players, such as Generative Adversarial Networks, multi-agent reinforcement learning, collaborative robots, and competing drones.
This course will use a framework encompassing all these problems called Variational Inequalities (VIs). Due to its generality, our studies also apply to standard minimization (single-player games). In addition to a brief introduction to game theory needed from a machine learning perspective, this course focuses on:
(i) Variational Inequalities and classes of VIs,
(ii) gradient-based optimization methods to solve VIs,
(iii) applications of VIs in machine learning problems.
Prerequisites: Linear algebra knowledge is required. Machine learning and convex optimization are recommended.
Grading Breakdown:
- 60% theoretical exam,
- 40% two projects in PyTorch (10% first project, 30% second project)
Homework and quizzes are not graded but part of the lecture content and solving them is strongly encouraged.
Only one exam will be offered in the exam session (this means you cannot improve your exam grade in a second attempt).
Date for lecture: Tuesdays, 14-16h.
- Students are encouraged to attend in person, as spontaneous discussions and brief lecture synopses will not be documented.
- We will provide a zoom link for online attendance. (As a courtesy, we will make a best effort attempt to publish recordings of some of the lectures. You should not assume that recordings will be available.)
First lecture: October 24, in CISPA lecture hall. E9.1 (CISPA building), room 0.05 (main lecture room)
Exercise sessions
We offer two slots (that alternate weekly - the detail schedule will be announced):
- Tuesday, 16-17h, in the lecture hall
- Thursday, 14-15h, on zoom
Registered students will find more information under the Lecture Outline tab.
Materials are posted under the materials tab.
About the lecturers:
- Tatjana Chavdarova is an incoming faculty at CISPA, with a research focus on the intersection of game theory and machine learning.
- Sebastian Stich is a faculty at CISPA, with a research focus on optimization for machine learning.