TUM Logo

Decentralized Learning in Untrusted Environments

Decentralized Learning in Untrusted Environments

Supervisor(s): Dr. Konstantin Boettinger, Daniel Kowatsch
Status: finished
Topic: Others
Author: Hyein Koo
Submission: 2023-03-15
Type of Thesis: Masterthesis
Thesis topic in co-operation with the Fraunhofer Institute for Applied and Integrated Security AISEC, Garching

Description

Federated learning has emerged as a privacy-preserving method for training a machine

learning model. Training a model in federated learning is performed by clients that

have local data, and a server manages the training process. The local data may

contain sensitive information that should not be shared to third parties. In this regard,

federated learning has been considered a secure training method that protects clients’

data. However, in recent studies it was shown that the clients’ data can be reconstructed

from their gradient updates sent to the server for aggregation.

We present a secure and scalable protocol for aggregation, random gradient mixing.

By exchanging gradient updates between clients before they are sent to the server, the

protocol enables anonymization and obfuscation of reconstructed data. The random

gradient mixing protocol provides security not only against a malicious server, but also

against honest-but-curious clients.

The security level of random gradient mixing depends on the hyperparameters. We

present the experimental result and evaluate the quality of reconstructions and the

performance of the models with different hyperparameters for random gradient mixing.

The choice of hyperparameters also affects the robustness of the random gradient

mixing protocol against clients dropping out. With a small range where a scale vector

is sampled from, it can be perfectly robust to dropouts. In order to mitigate potential

training failure due to dropouts, we also discuss an additional recovery phase.

The protocol is also efficient in communication. Since the communication cost

increases linearly with the number of clients, the protocol is expected to have a strong

scalability in practice.