LJPCheck: Functional Tests for Legal Judgment Prediction
Yuan Zhang, Wanhong Huang, Yi Feng, Chuanyi Li, Zhiwei Fei, Jidong Ge, Bin Luo, Vincent Ng
Abstract
Legal Judgment Prediction (LJP) refers to the task of automatically predicting judgment results (e.g., charges, law articles and term of penalty) given the fact description of cases. While SOTA models have achieved high accuracy and F1 scores on public datasets, existing datasets fail to evaluate specific aspects of these models (e.g., legal fairness, which significantly impact their applications in real scenarios). Inspired by functional testing in software engineering, we introduce LJPCHECK, a suite of functional tests for LJP models, to comprehend LJP models’ behaviors and offer diagnostic insights. We illustrate the utility of LJPCHECK on five SOTA LJP models. Extensive experiments reveal vulnerabilities in these models, prompting an in-depth discussion into the underlying reasons of their shortcomings.- Anthology ID:
- 2024.findings-acl.350
- Volume:
- Findings of the Association for Computational Linguistics: ACL 2024
- Month:
- August
- Year:
- 2024
- Address:
- Bangkok, Thailand
- Editors:
- Lun-Wei Ku, Andre Martins, Vivek Srikumar
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 5878–5894
- Language:
- URL:
- https://preview.aclanthology.org/fix-sig-urls/2024.findings-acl.350/
- DOI:
- 10.18653/v1/2024.findings-acl.350
- Cite (ACL):
- Yuan Zhang, Wanhong Huang, Yi Feng, Chuanyi Li, Zhiwei Fei, Jidong Ge, Bin Luo, and Vincent Ng. 2024. LJPCheck: Functional Tests for Legal Judgment Prediction. In Findings of the Association for Computational Linguistics: ACL 2024, pages 5878–5894, Bangkok, Thailand. Association for Computational Linguistics.
- Cite (Informal):
- LJPCheck: Functional Tests for Legal Judgment Prediction (Zhang et al., Findings 2024)
- PDF:
- https://preview.aclanthology.org/fix-sig-urls/2024.findings-acl.350.pdf