MPAF Model Poisoning Attacks To Federated Learning Based On Fake Clients
MPAF Model Poisoning Attacks To Federated Learning Based On Fake Clients
test accuracy
test accuracy
MPAF
0.6 0.6 0.6
test accuracy
test accuracy
MPAF
0.6 0.6 0.6
test accuracy
test accuracy
MPAF
0.6 0.6 0.6
Figure 2. Test accuracy of the global models learnt by different FL methods under the three attacks when the fraction of fake clients varies.
The datasets are MNIST (first row), Fashion-MNIST (second row) and Purchase (third row).
of the learnt global models as our metric. A lower test ac- model learnt with Trimmed-mean by 32% when there are
curacy indicates a stronger attack. 10% fake clients, while the baseline attacks can only de-
crease the test accuracy by at most 4%. Moreover, we also
5.2. Evaluation Results observe that MPAF is more effective when the fraction of
Impact of the fraction of fake clients: We explore the fake clients is larger. For instance, on Purchase dataset
impact of the fraction of fake clients on two baseline at- when Trimmed-mean is used, the test accuracy that MPAF
tacks (i.e., random attack and history attack) and MPAF. can reduce increases from 32% to 49% when the fraction of
Figure 2 shows the test accuracy of the global models learnt malicious clients increases from 10% to 25%.
by different FL methods when the fraction of fake clients Impact of the sample rate β: We evaluate the effective-
varies on the three datasets. We observe that when Fe- ness of MPAF when the server samples different fractions
dAvg is used, both baseline attacks and MPAF can reduce of clients in each FL round. Figure 3 shows the test accu-
the test accuracy of the learnt global models to random racy of the global models learnt with Trimmed-mean on all
guessing with only 1% fake clients. However, when clas- three datasets. We omit the results of non-robust FedAvg
sical defenses (e.g., Median and Trimmed-mean) are ap- for simplicity as the test accuracy is consistently close to
plied, MPAF can still significantly decrease the test accu- random guessing under MPAF. We observe that the sam-
racy while the baseline attacks cannot. For instance, on Pur- ple rate β does not have much impact on MPAF and that
chase dataset, MPAF reduces the test accuracy of the global MPAF can significantly decrease the test accuracy when β
1.0 1.0 1.0
test accuracy
test accuracy
0.6 0.6 0.6
Figure 3. Impact of the sample rate β on the test accuracy of the global models learnt by Trimmed-mean.
test accuracy
test accuracy
0.6 0.6 0.6
Figure 4. Impact of the scaling factor λ on the test accuracy of the global models learnt by Trimmed-mean.
ranges from 0.01 to 1.00. The previous claim that FedAvg main unchanged. Formally, a local model update g becomes
g
and classical defenses are robust to untargeted model poi- max(1,kgk2 /M ) after norm clipping. The largest `2 -norm of
soning attacks when β is small [21] does not apply to our the clipped local model updates is M . Therefore, the im-
attack. This is because their claim is based on the assump- pact of the malicious local model updates will be limited.
tion that an attacker can only compromise a small fraction As a result, the backdoor attacks [8] that rely on scaled lo-
of genuine clients. cal model updates will have lower attack success rate when
Impact of the scaling factor λ: We explore the impact of norm clipping is adopted as a countermeasure.
the scaling factor on MPAF. Figure 4 shows the test accu- We note that the idea of using norm clipping as a coun-
racy of the global models learnt by Trimmed-mean on all termeasure is not limited to backdoor attacks. In fact, it may
three datasets. We observe that the test accuracy first de- also be leveraged as a countermeasure against untargeted at-
creases as λ increases, and then remains almost unchanged tacks that involve scaling. In MPAF, we use a scaling factor
when λ further increases. Our results show that even though λ to increase the impact of fake local model updates during
the attacker does not know the hyperparameters of FL (e.g., aggregation. Therefore, it is intuitive to apply norm clipping
the global learning rate η), by choosing a reasonably large as a countermeasure against MPAF. We empirically evalu-
value for λ, e.g., λ ≥ 1 in our experiments, MPAF can re- ate the effectiveness of MPAF when norm clipping is used
duce the test accuracy of the global model significantly. as a countermeasure. Specifically, we use our default set-
ting for Fashion-MNIST dataset and Trimmed-mean as the
6. Norm Clipping as A Countermeasure aggregation rule. Before using Trimmed-mean to aggregate
A recent work [23] has proposed norm clipping as a the local model updates, we clip them with norm threshold
countermeasure against backdoor attacks in federated learn- M , where we vary the value of M in our experiments. We
ing. Specifically, the server selects a norm threshold M , omit the results of FedAvg for simplicity as the test accu-
and clips all local model updates whose `2 -norm is larger racy is consistently close to random guessing under MPAF.
than M such that their `2 -norm becomes M . The local Figure 5 shows the test accuracy of the global model
model updates whose `2 -norms are no larger than M re- learnt by Trimmed-mean on Fashion-MNIST. We use M →
based on fake clients. For instance, an interesting future
1.0 work is to improve MPAF with extra knowledge, e.g., train-
no attack ing data/model obtained from a similar learning task.
0.8 MPAF
test accuracy