Enhancing AMR Parsing with Group Relative Policy Optimization

Botond Barta, Endre Hamerlik, Milán Nyist, Masato Ito, Judit Acs


Abstract
We investigate the capabilities of the openly available Llama 3.2 1B language model for Abstract Meaning Representation (AMR) parsing through supervised fine-tuning, further enhanced by reinforcement learning via Group Relative Policy Optimization (GRPO). Existing supervised methods for AMR parsing face limitations due to static loss functions and challenges in capturing complex semantic phenomena. To address this, our GRPO-based approach explicitly optimizes fine-grained semantic rewards, including Smatch scores, frame-argument correctness, and structural validity of logical operations. Experimental results show that supervised fine-tuning alone establishes Llama as a capable English AMR parser, and subsequent GRPO fine-tuning further improves its performance. Our final model achieves higher Smatch scores, consistently respects critical low-level semantic constraints, and outperforms existing parsers on high-level semantic evaluation metrics across diverse linguistic phenomena.
Anthology ID:
2025.xllm-1.11
Volume:
Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Hao Fei, Kewei Tu, Yuhui Zhang, Xiang Hu, Wenjuan Han, Zixia Jia, Zilong Zheng, Yixin Cao, Meishan Zhang, Wei Lu, N. Siddharth, Lilja Øvrelid, Nianwen Xue, Yue Zhang
Venues:
XLLM | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
99–105
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.xllm-1.11/
DOI:
Bibkey:
Cite (ACL):
Botond Barta, Endre Hamerlik, Milán Nyist, Masato Ito, and Judit Acs. 2025. Enhancing AMR Parsing with Group Relative Policy Optimization. In Proceedings of the 1st Joint Workshop on Large Language Models and Structure Modeling (XLLM 2025), pages 99–105, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Enhancing AMR Parsing with Group Relative Policy Optimization (Barta et al., XLLM 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.xllm-1.11.pdf