Zhengwei Lv
2020
面向垂直领域的阅读理解数据增强方法(Method for reading comprehension data enhancement in vertical field)
Zhengwei Lv (吕政伟)
|
Lei Yang (杨雷)
|
Zhizhong Shi (石智中)
|
Xiao Liang (梁霄)
|
Tao Lei (雷涛)
|
Duoxing Liu (刘多星)
Proceedings of the 19th Chinese National Conference on Computational Linguistics
阅读理解问答系统是利用语义理解等自然语言处理技术,根据输入问题,对非结构化文档数据进行分析,生成一个答案,具有很高的研究和应用价值。在垂直领域应用过程中,阅读理解问答数据标注成本高且用户问题表达复杂多样,使得阅读理解问答系统准确率低、鲁棒性差。针对这一问题,本文提出一种面向垂直领域的阅读理解问答数据的增强方法,该方法基于真实用户问题,构造阅读理解训练数据,一方面降低标注成本,另一方面增加训练数据多样性,提升模型的准确率和鲁棒性。本文用汽车领域数据对该方法进行实验验证,其结果表明该方法对垂直领域阅读理解模型的准确率和鲁棒性均能有效提升。
2019
AUTOHOME-ORCA at SemEval-2019 Task 8: Application of BERT for Fact-Checking in Community Forums
Zhengwei Lv
|
Duoxing Liu
|
Haifeng Sun
|
Xiao Liang
|
Tao Lei
|
Zhizhong Shi
|
Feng Zhu
|
Lei Yang
Proceedings of the 13th International Workshop on Semantic Evaluation
Fact checking is an important task for maintaining high quality posts and improving user experience in Community Question Answering forums. Therefore, the SemEval-2019 task 8 is aimed to identify factual question (subtask A) and detect true factual information from corresponding answers (subtask B). In order to address this task, we propose a system based on the BERT model with meta information of questions. For the subtask A, the outputs of fine-tuned BERT classification model are combined with the feature of length of questions to boost the performance. For the subtask B, the predictions of several variants of BERT model encoding the meta information are combined to create an ensemble model. Our system achieved competitive results with an accuracy of 0.82 in the subtask A and 0.83 in the subtask B. The experimental results validate the effectiveness of our system.
Search
Co-authors
- Lei Yang 2
- Zhizhong Shi (石智中) 2
- Xiao Liang 2
- Tao Lei 2
- Duoxing Liu (刘多星) 2
- show all...