Stealing Knowledge from Pre-trained Language Models for Federated Classifier Debiasing
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review
Author(s)
Related Research Unit(s)
Detail(s)
Original language | English |
---|---|
Title of host publication | Medical Image Computing and Computer Assisted Intervention – MICCAI 2024 |
Subtitle of host publication | 27th International Conference, Marrakesh, Morocco, October 6–10, 2024, Proceedings, Part X |
Editors | Marius George Linguraru, Qi Dou, Aasa Feragen, Stamatia Giannarou, Ben Glocker, Karim Lekadir, Julia A. Schnabel |
Publisher | Springer |
Pages | 685-695 |
ISBN (electronic) | 978-3-031-72117-5 |
ISBN (print) | 978-3-031-72116-8 |
Publication status | Published - 3 Oct 2024 |
Publication series
Name | Lecture Notes in Computer Science |
---|
Conference
Title | 27th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2024) |
---|---|
Place | Morocco |
City | Marrakesh |
Period | 6 - 10 October 2024 |
Link(s)
DOI | DOI |
---|---|
Permanent Link | https://scholars.cityu.edu.hk/en/publications/publication(c7f2453f-31dc-4836-a3c2-4afff7efd698).html |
Abstract
Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that heterogeneous data of clients causes biased classifiers of local models during training, leading to the performance degradation of a federation system. In experiments, we surprisingly found that continuously freezing local classifiers can significantly improve the performance of the baseline FL method (FedAvg) for heterogeneous data. This observation motivates us to pre-construct a high-quality initial classifier for local models and freeze it during local training to avoid classifier biases. With this insight, we propose a novel approach named Federated Classifier deBiasing (FedCB) to solve the classifier biases problem in heterogeneous federated learning. The core idea behind FedCB is to exploit linguistic knowledge from pre-trained language models (PLMs) to construct high-quality local classifiers. Specifically, FedCB first collects the class concepts from clients and then uses a set of prompts to contextualize them, yielding language descriptions of these concepts. These descriptions are fed into a pre-trained language model to obtain their text embeddings. The generated embeddings are sent to clients to estimate the distribution of each category in the semantic space. Regarding these distributions as the local classifiers, we perform the alignment between the image representations and the corresponding semantic distribution by minimizing an upper bound of the expected cross-entropy loss. Extensive experiments on public datasets demonstrate the superior performance of FedCB compared to state-of-the-art methods. The source code is available at https://github.com/CUHK-AIM-Group/FedCB. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024
Research Area(s)
- Federated learning, Medical Image Classification, Pre-trained Language Model
Citation Format(s)
Stealing Knowledge from Pre-trained Language Models for Federated Classifier Debiasing. / Zhu, Meilu; Yang, Qiushi; Gao, Zhifan et al.
Medical Image Computing and Computer Assisted Intervention – MICCAI 2024: 27th International Conference, Marrakesh, Morocco, October 6–10, 2024, Proceedings, Part X. ed. / Marius George Linguraru; Qi Dou; Aasa Feragen; Stamatia Giannarou; Ben Glocker; Karim Lekadir; Julia A. Schnabel. Springer , 2024. p. 685-695 (Lecture Notes in Computer Science).
Medical Image Computing and Computer Assisted Intervention – MICCAI 2024: 27th International Conference, Marrakesh, Morocco, October 6–10, 2024, Proceedings, Part X. ed. / Marius George Linguraru; Qi Dou; Aasa Feragen; Stamatia Giannarou; Ben Glocker; Karim Lekadir; Julia A. Schnabel. Springer , 2024. p. 685-695 (Lecture Notes in Computer Science).
Research output: Chapters, Conference Papers, Creative and Literary Works › RGC 32 - Refereed conference paper (with host publication) › peer-review