Skip to main content

Research Repository

Advanced Search

FedBT: Effective and Robust Federated Unlearning via Bad Teacher Distillation for Secure Internet of Things

Wang, Fangwei; Huo, Jiashuai; Wang, Wei; Zhang, Xi; Liu, Yan; Tan, Zhiyuan; Wang, Changguang

Authors

Fangwei Wang

Jiashuai Huo

Wei Wang

Xi Zhang

Yan Liu

Changguang Wang



Abstract

Smart Internet of Things (IoT) devices generate vast, distributed data, and their limited computational and storage capacities complicate data protection. Federated Learning (FL) enables collaborative model training across clients, enhancing performance and protecting data privacy. The Right to be Forgotten (RTBF) raises the demand for precise data removal. Federated Unlearning (FU) offers a solution for accurate data deletion in FL systems. Existing FU methods often struggle to simultaneously ensure effective data forgetting and preserve model generalization. To mitigate these challenges, an effective and robust FU framework has been proposed, which is based on the “Bad Teacher” knowledge distillation (KD), termed FedBT. First, the “Bad Teacher" KD guides the trained model to eliminate specific client contributions from the global model. Next, the frequency domain extracts the global model’s generalization components. Finally, orthogonal constraints are applied to the KD-generated gradients within the orthogonal subspace of these components, ensuring the gradients preserve the trained model’s generalization ability. FedBT eliminates the need to store historical records of parameter updates. Using orthogonal space constraints, the generalization ability of the trained model is safeguarded during unlearning. Extensive experiments on three datasets with various metrics show our method reduces accuracy by only 0.53% on MNIST, 0.26% on Fashion-MNIST, and 4.67% on CIFAR10, surpassing the best approach. Furthermore, FedBT obtains an unlearning performance that most closely approximates the results obtained from retraining from scratch. FedBT boosts IoT security by enabling the “forgetting" of certain client data, crucial for protecting user privacy and ensuring secure device interactions.

Citation

Wang, F., Huo, J., Wang, W., Zhang, X., Liu, Y., Tan, Z., & Wang, C. (2025). FedBT: Effective and Robust Federated Unlearning via Bad Teacher Distillation for Secure Internet of Things. IEEE Internet of Things Journal, 12(15), 30634 - 30648. https://doi.org/10.1109/JIOT.2025.3571432

Journal Article Type Article
Acceptance Date May 15, 2025
Online Publication Date May 19, 2025
Publication Date Aug 1, 2025
Deposit Date May 15, 2025
Publicly Available Date Jul 2, 2025
Journal IEEE Internet of Things Journal
Electronic ISSN 2327-4662
Publisher Institute of Electrical and Electronics Engineers
Peer Reviewed Peer Reviewed
Volume 12
Issue 15
Pages 30634 - 30648
DOI https://doi.org/10.1109/JIOT.2025.3571432
Keywords Federated unlearning, bad teacher, knowledge distillation, frequency domain, orthogonal subspace
Public URL http://researchrepository.napier.ac.uk/Output/4292075
This output contributes to the following UN Sustainable Development Goals:

SDG 9 - Industry, Innovation and Infrastructure

Build resilient infrastructure, promote inclusive and sustainable industrialisation and foster innovation

Files

FedBT: Effective And Robust Federated Unlearning Via Bad Teacher Distillation For Secure Internet Of Things (accepted version) (3.8 Mb)
PDF








You might also like



Downloadable Citations