@inproceedings{qin-etal-2025-federated,
title = "Federated Data-Efficient Instruction Tuning for Large Language Models",
author = "Qin, Zhen and
Wu, Zhaomin and
He, Bingsheng and
Deng, Shuiguang",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.803/",
pages = "15550--15568",
ISBN = "979-8-89176-256-5",
abstract = "Instruction tuning is a crucial step in improving the responsiveness of pretrained large language models (LLMs) to human instructions. Federated learning (FL) helps to exploit the use of vast private instruction data from clients, becoming popular for LLM tuning by improving data diversity. Existing federated tuning simply consumes all local data, causing excessive computational overhead and overfitting to local data, while centralized data-efficient solutions are not suitable for FL due to privacy concerns. This work presents FedHDS, a federated data-efficient instruction tuning approach, which tunes LLMs with a representative subset of edge-side data. It reduces the data redundancy at both intra- and inter-client levels without sharing raw data. Experiments with various LLMs, datasets and partitions show that FedHDS improves Rouge-L on unseen tasks by an average of 10.72{\%} over the SOTA full-data federated instruction tuning methods, while using less than 1.5{\%} of the data samples, improving training efficiency by up to tens of times."
}
Markdown (Informal)
[Federated Data-Efficient Instruction Tuning for Large Language Models](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.803/) (Qin et al., Findings 2025)
ACL