GLGE: A New General Language Generation Evaluation Benchmark
Dayiheng Liu, Yu Yan, Yeyun Gong, Weizhen Qi, Hang Zhang, Jian Jiao, Weizhu Chen, Jie Fu, Linjun Shou, Ming Gong, Pengcheng Wang, Jiusheng Chen, Daxin Jiang, Jiancheng Lv, Ruofei Zhang, Winnie Wu, Ming Zhou, Nan Duan
- Anthology ID:
- 2021.findings-acl.36
- Volume:
- Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
- Month:
- August
- Year:
- 2021
- Address:
- Online
- Venue:
- Findings
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 408–420
- Language:
- URL:
- https://aclanthology.org/2021.findings-acl.36
- DOI:
- 10.18653/v1/2021.findings-acl.36
- Cite (ACL):
- Dayiheng Liu, Yu Yan, Yeyun Gong, Weizhen Qi, Hang Zhang, Jian Jiao, Weizhu Chen, Jie Fu, Linjun Shou, Ming Gong, Pengcheng Wang, Jiusheng Chen, Daxin Jiang, Jiancheng Lv, Ruofei Zhang, Winnie Wu, Ming Zhou, and Nan Duan. 2021. GLGE: A New General Language Generation Evaluation Benchmark. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 408–420, Online. Association for Computational Linguistics.
- Cite (Informal):
- GLGE: A New General Language Generation Evaluation Benchmark (Liu et al., Findings 2021)
- PDF:
- https://preview.aclanthology.org/ingestion-script-update/2021.findings-acl.36.pdf
- Code
- microsoft/glge
- Data
- GLGE, BookCorpus, CLUE, CoQA, FLUE, GLUE, IndoNLU Benchmark, SQuAD, SuperGLUE