News
Additional Material regarding Privacy I on the vector norms and the proof of the Laplace MechanismWritten on 01.05.25 by Adam Dziedzic Hello Everyone, We uploaded the additional slides on the vector norms and the proof of the Laplace Mechanism. Please, find them here: https://cms.cispa.saarland/tml2025/materials/index Adam & Franziska |
Assignment 1Written on 30.04.25 by Adam Dziedzic Hello Everyone,
Hello Everyone,
Good luck! |
Team Registration open until May 5thWritten on 29.04.25 (last change on 29.04.25) by Franziska Boenisch Hey everyone,
the team registration has been opened for the programming assignments. Please submit your team of 2 people until May 5th. We will then distribute the access tokens to our score board to the teams, such that you'll be able to submit your assignments. Without a token, you… Read more Hey everyone,
the team registration has been opened for the programming assignments. Please submit your team of 2 people until May 5th. We will then distribute the access tokens to our score board to the teams, such that you'll be able to submit your assignments. Without a token, you won't be able to submit the assignment, hence, it is crucial that you register your team in time!
Kind regards Franziska and Adam |
Trustworthy Machine Learning
Organization
Time: Wednesdays from 14:00 to 16:00.
Location: CISPA Building C0 (Stuhlsatzenhaus 5 66123 Saarbrücken) Lecture Hall Ground Floor (Room 0.05)
Starting Date: April 16th, 2025
Description: The deployment of machine learning applications in real-world systems necessitates methods to ensure their trustworthiness. This course explores the different aspects of trustworthy machine learning, including Privacy, Collaborative Learning, Model Confidentiality, Robustness, Fairness and Bias, Explainability, Security, and Governance.
Learning Objectives: The objective of this tutorial is to provide attendees with a comprehensive understanding of trustworthy machine learning, covering key aspects such as privacy, robustness, fairness, explainability, security, and governance. Participants will benefit by gaining practical skills in identifying and mitigating risks associated with machine learning models, including privacy attacks, model theft, bias, and adversarial threats. By the end of the course, attendees are expected to have enhanced their knowledge of cutting-edge defense strategies, developed practical skills in securing machine learning systems, and deepened their understanding of the ethical and societal implications of deploying these models in real-world scenarios.
What you have to do weekly?
This is a flipped classroom lecture. What does this mean?
You have to watch the lecture videos a week before the lecture. For example, if we consider "Lecture 1, Privacy" on April 30th in class, you should watch the video on April 23rd. Then, if you have questions regarding the lecture, you can post them in the forum. If you want your question to be considered for an answer in the in-person session (in that example on April 30th), you should ask them until the Friday before, i.e., in the example until April 25th. Questions regarding the course content will not be individually answered by the instructors in the forum. They will be answered in the in-person lecture hours. The in-person lecture hours are not recorded. Hence, if you have questions, you need to post them on time, and attend the Q&A session.
Additionally, we will post ungraded theoretical questions regarding the individual lectures. They give you an impression on how the exam and midterm will look like. We highly suggest solving them. Solutions can also be discussed in the in-person Q&A session.
What are our topics?
Privacy: We will analyze the landscape of privacy attacks against machine learning models and study the means to prevent these attacks.
Collaborative Learning: We will analyze the risks to trustworthy machine learning that arise in collaborative machine learning setups and look into their mitigations.
Model Confidentiality: We will see that machine learning models can be easily stolen through different methods, such as simple copies of the models, or the private training data, or the extraction of the model exposed via a public or private API. We will analyze different attack strategies to steal the models and the state-of-the-art defense methods.
Robustness: We will learn about different facets of robustness, such as robustness to out-of-distribution samples, natural noise present in the input data, or adversarial examples, where attackers incur imperceptible changes to the input to ML models to fool their predictions.
Fairness and Bias: We will scrutinize the behavior of ML models on different subgroups of the training data. We will assess the models’ responses to inputs with different attributes and will uncover the potential causes of unfair or biased model responses and current mitigations.
Explainability: We will address the challenges that arise from machine learning models’ black-box behavior and look into techniques to explain predictive behavior.
Security and Governance: Machine learning applications are usually not isolated but integrated in some systems in a given society with their respective norms and values. We will study the security risks that can arise from the deployment of machine learning applications and how to use governance approaches to mitigate these risks.
Throughout the course, we will discuss outstanding challenges and future research directions to make machine learning more robust, private, and trustworthy.
Assignments (with tentative dates):
The course entails 4 practical graded assignments based on implementing the concepts studied during the lecture.
Assignments need to be handed in groups of two. If you do not have a partner, yet, you can use the forum to find one.
-
Privacy: Implement a membership inference attack and achieve the highest attack success. We give a model and a list of data points, your task is to determine for each point yes/no, meaning: member/no-member. --> Due 28.5.
-
Model extraction: We will offer a machine-learning model over an API and you will extract the model behavior. You upload a PyTorch model. We check how close the uploaded model is in terms of the predictions to the victim model. --> Due 11.6.
-
Robustness: Train a model as robust as possible against the highest number of adversarial examples we will generate. You submit a PyTorch model, we load it, and run the attack. The goal is to create a model with the highest adversarial and clean accuracy. --> Due 2.7.
-
Fairness: Train a classifier that has the highest demographic parity. You submit a PyTorch model and we assess how fair it is. --> Due 30.7.
Submission of Assignments:
1. Make a GitHub repository.
2. If it is private, add Adam and Franziska (adam-dziedzic and fraboeni).
3. For your final submission, please make a tag (click on tags, "create new release", create it)
4. Submit the link to the tagged version as the online Test in CMS under Assignment X (X is the number of the current assignment).
Exam and Grading
Exam format: written.
Tentative Lecture Dates
The videos for each lecture are linked at the respective lecture date. To access the full playlist, visit: https://www.youtube.com/playlist?list=PLNfU-a7sxIwvS7dhnOPdFtvhdNcrnufEW
-
Overview, Administration & Intro (16.4.)
-
Model stealing and defenses (21.5. - supervised learning (SL), - 28.5. self-supervised learning (SSL).)
-
Midterm-Exam 4.6. (We'll first be writing the exam, and then covering the questions about the robustness lecture, which is not going to be part of the midterm)
-
Adversarial Machine Learning / Robustness (4.6.)
-
Fairness and bias (25.6.)
-
Explainability (2.7.)
-
Security and Governance (9.7.)
-
Summary & Open Questions (9.7.)
-
Final Exam on 30.7 (30th of July) from 16.00 - 18.00. Location: lecture hall 002 in E1 3 building on UdS campus.
Do I have the right qualifications for the lecture?
Students are required to have successfully completed a basic lecture on machine learning.
Additionally, you need to have hands-on coding experience with Python for Machine Learning. This means, preferably PyTorch (alternatively TensorFlow). All assignments will require you to do this type of coding. If you are unfamiliar and have never coded in PyTorch, it will be a steep learning curve.