@inproceedings{tang-tang-2025-system,
title = "System Report for {CCL}25-Eval Task 5: Hierarchical Multi-Task Prompt Fine-Tuning and {PPO} Reinforcement for Classical {C}hinese Poetry Comprehension and Sentiment Reasoning",
author = "Tang, Jingjun and
Tang, Zhiwen",
editor = "Lin, Hongfei and
Li, Bin and
Tan, Hongye",
booktitle = "Proceedings of the 24th {C}hina National Conference on Computational Linguistics ({CCL} 2025)",
month = aug,
year = "2025",
address = "Jinan, China",
publisher = "Chinese Information Processing Society of China",
url = "https://preview.aclanthology.org/ingest-ccl/2025.ccl-2.23/",
pages = "192--199",
abstract = "``We present a hierarchical multi-task framework to enhance classical Chinese poetry understand-ing and sentiment reasoning using large language models. Centered on Qwen2.5-14B-Instruction or Xunzi-Qwen-14B, we construct a 1,225-sample corpus of Tang and Song poems with parallel translations and multi-label sentiment annotations (e.g., nostalgia, patriotism, contemplation).The task is divided into comprehension, translation, and sentiment inference, each guided by dynamic prompting and task-specific templates. We employ mixed supervised fine-tuning to better capture syntactic and metaphorical patterns. For sentiment reasoning, we apply proximal policy optimization (PPO) with a custom reward function, boosting accuracy from 0.771 to 0.807(p {\ensuremath{<}} 0.01). Our model achieves a 0.714 comprehensive score, outperforming single-task base-lines by 12.6{\%}. Ablation studies further confirm the benefits of multi-task learning in promoting cross-task knowledge transfer.Keywords: Classical Chinese Poetry, Multi-Task Fine-Tuning, Data Augmentation, ProximalPolicy Optimization''"
}Markdown (Informal)
[System Report for CCL25-Eval Task 5: Hierarchical Multi-Task Prompt Fine-Tuning and PPO Reinforcement for Classical Chinese Poetry Comprehension and Sentiment Reasoning](https://preview.aclanthology.org/ingest-ccl/2025.ccl-2.23/) (Tang & Tang, CCL 2025)
ACL