Stealing Knowledge from Pre-trained Language Models for Federated Classifier Debiasing

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review

View graph of relations

Author(s)

Detail(s)

Original languageEnglish
Title of host publicationMedical Image Computing and Computer Assisted Intervention – MICCAI 2024
Subtitle of host publication27th International Conference, Marrakesh, Morocco, October 6–10, 2024, Proceedings, Part X
EditorsMarius George Linguraru, Qi Dou, Aasa Feragen, Stamatia Giannarou, Ben Glocker, Karim Lekadir, Julia A. Schnabel
PublisherSpringer 
Pages685-695
ISBN (electronic)978-3-031-72117-5
ISBN (print)978-3-031-72116-8
Publication statusPublished - 3 Oct 2024

Publication series

NameLecture Notes in Computer Science

Conference

Title27th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2024)
PlaceMorocco
CityMarrakesh
Period6 - 10 October 2024

Abstract

Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that heterogeneous data of clients causes biased classifiers of local models during training, leading to the performance degradation of a federation system. In experiments, we surprisingly found that continuously freezing local classifiers can significantly improve the performance of the baseline FL method (FedAvg) for heterogeneous data. This observation motivates us to pre-construct a high-quality initial classifier for local models and freeze it during local training to avoid classifier biases. With this insight, we propose a novel approach named Federated Classifier deBiasing (FedCB) to solve the classifier biases problem in heterogeneous federated learning. The core idea behind FedCB is to exploit linguistic knowledge from pre-trained language models (PLMs) to construct high-quality local classifiers. Specifically, FedCB first collects the class concepts from clients and then uses a set of prompts to contextualize them, yielding language descriptions of these concepts. These descriptions are fed into a pre-trained language model to obtain their text embeddings. The generated embeddings are sent to clients to estimate the distribution of each category in the semantic space. Regarding these distributions as the local classifiers, we perform the alignment between the image representations and the corresponding semantic distribution by minimizing an upper bound of the expected cross-entropy loss. Extensive experiments on public datasets demonstrate the superior performance of FedCB compared to state-of-the-art methods. The source code is available at https://github.com/CUHK-AIM-Group/FedCB. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2024

Research Area(s)

  • Federated learning, Medical Image Classification, Pre-trained Language Model

Citation Format(s)

Stealing Knowledge from Pre-trained Language Models for Federated Classifier Debiasing. / Zhu, Meilu; Yang, Qiushi; Gao, Zhifan et al.
Medical Image Computing and Computer Assisted Intervention – MICCAI 2024: 27th International Conference, Marrakesh, Morocco, October 6–10, 2024, Proceedings, Part X. ed. / Marius George Linguraru; Qi Dou; Aasa Feragen; Stamatia Giannarou; Ben Glocker; Karim Lekadir; Julia A. Schnabel. Springer , 2024. p. 685-695 (Lecture Notes in Computer Science).

Research output: Chapters, Conference Papers, Creative and Literary WorksRGC 32 - Refereed conference paper (with host publication)peer-review