Liu, I. JiehI. JiehLiuLin, Ci SiangCi SiangLinYang, Fu EnFu EnYangYU-CHIANG WANG2024-04-242024-04-242024-03-2521595399https://scholars.lib.ntu.edu.tw/handle/123456789/642022Federated Learning (FL) is an emerging paradigm that enables multiple users to collaboratively train a robust model in a privacy-preserving manner without sharing their private data. Most existing approaches of FL only consider traditional single-label image classification, ignoring the impact when transferring the task to multi-label image classification. Nevertheless, it is still challenging for FL to deal with user heterogeneity in their local data distribution in the real-world FL scenario, and this issue becomes even more severe in multi-label image classification. Inspired by the recent success of Transformers in centralized settings, we propose a novel FL framework for multi-label classification. Since partial label correlation may be observed by local clients during training, direct aggregation of locally updated models would not produce satisfactory performances. Thus, we propose a novel FL framework of Language-Guided Transformer (FedLGT) to tackle this challenging task, which aims to exploit and transfer knowledge across different clients for learning a robust global model. Through extensive experiments on various multi-label datasets (e.g., FLAIR, MS-COCO, etc.), we show that our FedLGT is able to achieve satisfactory performance and outperforms standard FL techniques under multi-label FL scenarios. Code is available at https://github.com/Jack24658735/FedLGT.[SDGs]SDG17Language-Guided Transformer for Federated Multi-Label Classificationconference paper10.1609/aaai.v38i12.292952-s2.0-85189522700https://api.elsevier.com/content/abstract/scopus_id/85189522700