M.Sc. Marvin Xhemrishi
Technical University of Munich
Associate Professorship of Coding and Cryptography (Prof. Wachter-Zeh)
Postal address
Theresienstr. 90
80333 München
- Phone: +49 (89) 289 - 29062
- Room: 0104.03.415A
- marvin.xhemrishi@tum.de
Theses in Progress
Differentially-Private and Robust Federated Learning
Description
Federated learning is a machine learning paradigm that aims to learn collaboratively from decentralized private data owned by entities referred to as clients. However, due to its decentralized nature, federated learning is susceptible to poisoning attacks, where malicious clients try to corrupt the learning process by modifying their data or local model updates. Moreover, the updates sent by the clients might leak information about the private data involved in the learning. This thesis aims to investigate and combine existing robust aggregation techniques in FL with differential privacy techniques.
References:
[1] - https://arxiv.org/pdf/2304.09762.pdf
[2] - https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9757841
[3] - https://dl.acm.org/doi/abs/10.1145/3465084.3467919
Prerequisites
- Knowledge about machine learning and gradient descent optimization
- Proficiency in Python and PyTorch
- Undergraduate statistics courses
- Prior knowledge about differential privacy is a plus
Contact
marvin.xhemrishi@tum.de
luis.massny@tum.de
Supervisor:
Secure Federated Learning with Differential Privacy
Description
Federated learning is a machine learning paradigm that aims to learn collaboratively from decentralized private data owned by entities referred to as clients. However, due to its decentralized nature, federated learning is susceptible to model poisoning attacks, where malicious clients try to corrupt the learning process by modifying local model updates. Moreover, the updates sent by the clients might leak information about the private data involved in the learning. The goal of this work is to investigate and combine existing robust aggregation techniques in FL with differential privacy techniques.
References:
[1] - https://arxiv.org/pdf/2304.09762.pdf
[2] - https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=9757841
[3] - https://dl.acm.org/doi/abs/10.1145/3465084.3467919
Prerequisites
- Basic knowledge about machine learning and gradient descent optimization
- First experience with machine learning in python
- Undergraduate statistics courses
- Prior knowledge about differential privacy is a plus
Supervisor:
Contribution scoring in Federated Learning
Description
Federated learning (FL) is a machine learning paradigm that aims to learn collaboratively from decentralized private data, owned by entities referred to as clients. In real-world applications of FL it is important to score the contribution of each client. The goal of this seminar is to provide a high-level overview of existing contribution-scoring techniques in federated learning using [1-2] and other references.
Prerequisites
References:
[1] - https://ieeexplore.ieee.org/document/10138056
[2] - https://arxiv.org/pdf/2403.07151.pdf
Supervisor:
Implementation of model poisoning attacks in federated learning
Description
Federated learning is a machine learning paradigm where decentralized entities (clients) collaboratively learn using their private data. A central server acts as a coordinator of the learning process. Due to the sensitivity of the private data involved, the data cannot be transferred. A salient problem of federated learning is the presence of malicious clients, which are clients that try to destroy the learning process. Malicious clients can do this by corrupting their data and/or by modifying their local model updates. The goal of this project is to understand how model poisoning attacks and defense strategies perform under different scenarios of federated learning using experiments.
References:
[1]- https://www.ndss-symposium.org/wp-content/uploads/ndss2021_6C-3_24498_paper.pdf
[2]- https://arxiv.org/pdf/1903.03936.pdf
[3]- https://arxiv.org/pdf/2304.00160.pdf
Prerequisites
- Basic knowledge of machine learning
- Python programming skills, knowledge of PyTorch is an advantage
Contact
marvin.xhemrishi@tum.de
Supervisor:
2022
- Adaptive private distributed matrix multiplication. IEEE Transactions on Information Theory, 2022 more…
- The Wiretap Channel for Capacitive PUF-Based Security Enclosures. IACR Transactions on Cryptographic Hardware and Embedded Systems, 2022 more…
- Analysis of Communication Channels Related to Physical Unclonable Functions. Workshop on Coding and Cryptography (WCC), 2022 more…
- Distributed Linear Computations over Private Sparse Matrices. IEEE European School of Information Theory, 2022 more…
- Distributed Linear Computations over Private Sparse Matrices. Munich Workshop on Coding and Cryptography, 2022 more…
- Computational Code-Based Privacy for Coded Federated Learning. TUM ICE Workshop Raitenhaslach, 2022 more…
- Distributed Matrix-Vector Multiplication with Sparsity and Privacy Guarantees. IEEE International Symposium on Information Theory, 2022 more…
- Efficient Private Storage of Sparse Machine Learning Data. IEEE Information Theory Workshop, 2022 more…
- Computational Code-Based Privacy in Coded Federated Learning. IEEE International Symposium on Information Theory, 2022 more…
2021
2020
2019
- List Decoding of Short Codes for Communication over Unknown Fading Channels. , Workshop on Coding, Cooperation, and Security in Modern Communication Networks (COCO 2019) . , 2019 more…
- List Decoding for Fading Channels. Oberpfaffenhofen Workshop on High Throughput Coding (OWHTC) , 2019 more…
- List Decoding of Short Codes for Communication over Unknown Fading Channels. Asilomar Conference on Signals, Systems, and Computers, 2019 more…