Multi-task Legal Judgement Prediction Combining a Subtask of Seriousness of Charge

Xu Zhuopeng, Li Xia, Li Yinlin, Wang Zihan, Fanxu Yujie, Lai Xiaoyan


Abstract
Legal Judgement Prediction has attracted more and more attention in recent years. One of the challenges is how to design a model with better interpretable prediction results. Previous studies have proposed different interpretable models based on the generation of court views and the extraction of charge keywords. Different from previous work, we propose a multi-task legal judgement prediction model which combines a subtask of the seriousness of charges. By introducing this subtask, our model can capture the attention weights of different terms of penalty corresponding to the charges and give more attention to the correct terms of penalty in the fact descriptions. Meanwhile, our model also incorporates the position of defendant making it capable of giving attention to the contextual information of the defendant. We carry several experiments on the public CAIL2018 dataset. Experimental results show that our model achieves better or comparable performance on three subtasks compared with the baseline models. Moreover, we also analyze the interpretable contribution of our model.
Anthology ID:
2020.ccl-1.105
Volume:
Proceedings of the 19th Chinese National Conference on Computational Linguistics
Month:
October
Year:
2020
Address:
Haikou, China
Venue:
CCL
SIG:
Publisher:
Chinese Information Processing Society of China
Note:
Pages:
1132–1142
Language:
English
URL:
https://aclanthology.org/2020.ccl-1.105
DOI:
Bibkey:
Cite (ACL):
Xu Zhuopeng, Li Xia, Li Yinlin, Wang Zihan, Fanxu Yujie, and Lai Xiaoyan. 2020. Multi-task Legal Judgement Prediction Combining a Subtask of Seriousness of Charge. In Proceedings of the 19th Chinese National Conference on Computational Linguistics, pages 1132–1142, Haikou, China. Chinese Information Processing Society of China.
Cite (Informal):
Multi-task Legal Judgement Prediction Combining a Subtask of Seriousness of Charge (Zhuopeng et al., CCL 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/update-css-js/2020.ccl-1.105.pdf