News

July 16 - Talk & Poster Session

Written on 09.07.24 (last change on 09.07.24) by Sebastian Stich

Dear students,

On July 16, we will again meet in room 0.05, and have a guest talk on personalization in federated learning from 16:15 to 17:15, followed by the poster presentations from 17:15 to 18:00.

Please send your poster by Monday morning (8am) to Yuan Gao (...) if you want us to print it… Read more

Dear students,

On July 16, we will again meet in room 0.05, and have a guest talk on personalization in federated learning from 16:15 to 17:15, followed by the poster presentations from 17:15 to 18:00.

Please send your poster by Monday morning (8am) to Yuan Gao (...) if you want us to print it for you. Alternatively, you can print it yourself (see lecture slides).

Feel free to share the invitation for the talk and the poster session with your friends and colleagues.

-------------------
Time and location: July 16, 16:15-17:15, CISPA C0 room 0.05

Title: Personalization Mitigates the Perils of Local SGD for Heterogeneous Distributed Learning

Abstract: Local SGD or Federated Averaging is one of the most popular algorithms for large-scale distributed optimization, such as cross-device federated learning. However, it has been challenging to prove its efficacy against simpler algorithms such as mini-batch SGD. In this talk, we will discuss the limitations of the Local SGD algorithm in even simple convex problems and motivate a personalized variant of Local SGD. We will discuss new convergence guarantees for this personalized approach, highlighting its dependence on existing notions of data heterogeneity, and compare these guarantees to Local SGD. Our theoretical analysis reveals that in scenarios with low data heterogeneity, personalized Local SGD outperforms both pure local training on a single machine and local SGD/mini-batch SGD that produce a consensus model across all machines. This performance gain arises because the personalized approach avoids the fixed point discrepancy due to its local updates and can reduce the consensus error between machines to zero, even with a constant learning rate. We support our findings with experiments on distributed linear regression tasks with varying degrees of data heterogeneity.

Bio: Kumar Kshitij Patel is a fifth-year PhD student at the Toyota Technological Institute at Chicago (TTIC), where Professors Nati Srebro and Lingxiao Wang advise him, and his research centers on optimization algorithms in practically relevant settings such as federated learning. He wants to understand the effects of limited communication budgets, data heterogeneity, sequential decision-making, and privacy considerations in such settings. To systematically understand these settings, Kshitij has characterized the min-max oracle complexity of optimization for several representative problem classes with varying heterogeneity. He has also explored the game-theoretic aspects of collaboration protocols to prevent the formation of inaccessible data silos and sustain collaborations despite strategic agent behavior as well as understand how heterogeneous distribution shifts can affect deployed machine learning models.  Additionally, he is interested in designing better and more practical privacy defenses that go beyond traditional differential privacy techniques for large machine learning models, such as diffusion models, to prevent them from memorizing sensitive data.  

During the summer of 2023, Kshitij worked with the privacy-preserving machine learning team at Sony AI alongside Lingjuan Lyu. In the summer of 2020, he interned as an applied scientist with the CodeGuru team at Amazon Web Services. Before joining TTIC, Kshitij obtained his BTech in Computer Science and Engineering from the Indian Institute of Technology, Kanpur, where working with Professor Purushottam Kar on Bandit Learning algorithms. He also spent a year on academic exchange at École Polytechnique Fédérale de Lausanne (EPFL), working in the Machine Learning and Optimization Laboratory (MLO) with Professor Martin Jaggi.

Tutorial this week

Written on 08.07.24 by Yuan Gao

Hi all,

Please note that this week we will resume the tutorial and discuss exercise sheet 10 (the final exercise sheet) at our usual times.

Best,

Yuan

Lecture updates

Written on 10.06.24 by Sebastian Stich

Dear Students

  • A reminder: this week, the lecture will take place online (on zoom) only. Next week we can meet at CISPA as usual.
  • The exam questions will be discussed this week (& next week) in the exercise sessions.
  • It would be very helpful if you could fill out this short survey regarding… Read more

Dear Students

  • A reminder: this week, the lecture will take place online (on zoom) only. Next week we can meet at CISPA as usual.
  • The exam questions will be discussed this week (& next week) in the exercise sessions.
  • It would be very helpful if you could fill out this short survey regarding the lecture format: https://forms.gle/xBSKAVPgsRBz4wSf8.
  • The exam grades have been released. I will bring the exams to the lecture on June 18 for inspection (alternatively, you can also request a copy by email).

Sebastian

Midterm Exam

Written on 07.05.24 by Sebastian Stich

Dear students

The midterm exam takes place on June 4 in the lecture slot (start: 16:15h, rooms 0.01 - 0.05 at CISPA). Please register on CMS if you want to take the exam.

(Note: a registration on LSF is not needed for midterms).

Additional Tutorial Session

Written on 29.04.24 by Sebastian Stich

In addition to the session on Tuesday afternoons, we will now offer an additional exercise session on Mondays from 13:15 to 14:00 on Zoom (starting next week).

This session is primarily intended for students who cannot attend the Tuesday session, but everyone is welcome to join.

The main goal of… Read more

In addition to the session on Tuesday afternoons, we will now offer an additional exercise session on Mondays from 13:15 to 14:00 on Zoom (starting next week).

This session is primarily intended for students who cannot attend the Tuesday session, but everyone is welcome to join.

The main goal of these exercise sessions is to discuss exercises, material from the lecture, or any other questions you may have. If there are no questions, the session may end early, so we encourage everyone to join on time.

First lecture today

Written on 16.04.24 (last change on 16.04.24) by Sebastian Stich

Welcome to the OPTML course 2024!

The first lecture will be today at 4.15pm in the CISPA building, room 0.01.

The lecture can also be followed on zoom: (link available to registered students).

Show all

Optimization for Machine Learning

This course teaches an overview of modern mathematical optimization methods for machine learning and data science applications. In particular, the scalability of algorithms to large datasets will be discussed in theory and in implementation.

This advanced lecture aims to prepare students to research this topic. An interest in and the ability to understand and apply mathematical proofs are essential.

The first lecture will take place on April 16 (hybrid), building E9 1 (CISPA), room 0.01 ("presentation"). The zoom link will be sent by email to registered participants.

 

Learning Prerequisites

  • Previous coursework in calculus, linear algebra, and probability is required.
  • Familiarity with optimization and/or machine learning is beneficial.

Students are recommended to register for this course only as master's students, but attendance is also possible for bachelor students in their last semester. There are no strict rules or regulations, but the students must acquire (missing) fundamentals independently. Please note that this course will be offered regularly (yearly).

 

Course Information 2024

  • The lectures will take place on Tuesday 4-6pm, with exercise sessions just before (2-4pm).

 

Program

Course materials will be posted under the materials tab, the schedule and suggested reading materials will be updated under the contents tab.

Preliminary schedule (subject to change):

  • April 16, first lecture
  • June 4, midterm exam
  • July 16, project presentation
  • July 23, no lecture
  • TBD (summer break), Additional Tutorial/Exam Q&A 
  • (TBC): September 3, afternoon, Final Exam

Note that only one exam will take place this year (re-exam possible after the next reading in 2025).

 

Registration

  • Students are required to register for the lecture and exam on LSF (the registration deadline is a few weeks after the first lecture and will be visible on LSF). No assistance can be provided if you miss this deadline.

 

Organization

  • There will be one lecture per week on Tuesdays, 16:15 - 17:45. The first lecture is on April 16.
  • There will be Q&A sessions with the teaching assistants
    • on Tuesdays, 15:00 - 16:00 (room 0.01). The first Q&A session is on April 23.
    • on Mondays, 13:15 - 14:00 (zoom). The first session is on May 6. 
    • Attendance of the tutorials/Q&A session is not mandatory but strongly recommended.
  • Office hours: before/after each lecture.
  • We will use a hybrid format. All lectures can be attended online, some lectures will be held in building E9 1 (CISPA), room 0.01 (presentation room, ground floor).

 

Grading and Exam

  • (requirement) A research project (theoretical or practical) will focus on either practical implementation and the real-world performance of one of the studied optimization algorithms or variants or a theoretical investigation with a small extension of one of these schemes. The project is mandatory and done in groups of 2-3 students. The projects will be graded as: Fail, Pass, Above Average (top 30%, 0.3 bonus), or Excellent (top 10%, 0.6 bonus). You are required to pass the project to take part in the exam. Failed projects can be resubmitted within 2 weeks.
  • (25%) Points from the midterm exam will count for 25% of the final mark.
  • (75%) Points from the final exam will count for 75% of the final mark.
  • The exams will be written and cover all the material discussed in the lectures and the topics from the assignments/exercises. 
  • If you pass the course (at least 4.0), eventual bonus points from the project will be considered to determine the final grade.

 

Privacy Policy | Legal Notice
If you encounter technical problems, please contact the administrators.