@inproceedings{anni-etal-2022-eventbert,
    title = "{E}vent{BERT}: Incorporating Event-based Semantics for Natural Language Understanding",
    author = "Anni, Zou  and
      Zhuosheng, Zhang  and
      Hai, Zhao",
    editor = "Sun, Maosong  and
      Liu, Yang  and
      Che, Wanxiang  and
      Feng, Yang  and
      Qiu, Xipeng  and
      Rao, Gaoqi  and
      Chen, Yubo",
    booktitle = "Proceedings of the 21st Chinese National Conference on Computational Linguistics",
    month = oct,
    year = "2022",
    address = "Nanchang, China",
    publisher = "Chinese Information Processing Society of China",
    url = "https://preview.aclanthology.org/ingest-emnlp/2022.ccl-1.69/",
    pages = "774--785",
    language = "eng",
    abstract = "``Natural language understanding tasks require a comprehensive understanding of natural language and further reasoning about it, on the basis of holistic information at different levels to gain comprehensive knowledge. In recent years, pre-trained language models (PrLMs) have shown impressive performance in natural language understanding. However, they rely mainly on extracting context-sensitive statistical patterns without explicitly modeling linguistic information, such as semantic relationships entailed in natural language. In this work, we propose EventBERT, an event-based semantic representation model that takes BERT as the backbone and refines with event-based structural semantics in terms of graph convolution networks. EventBERT benefits simultaneously from rich event-based structures embodied in the graph and contextual semantics learned in pre-trained model BERT. Experimental results on the GLUE benchmark show that the proposed model consistently outperforms the baseline model.''"
}Markdown (Informal)
[EventBERT: Incorporating Event-based Semantics for Natural Language Understanding](https://preview.aclanthology.org/ingest-emnlp/2022.ccl-1.69/) (Anni et al., CCL 2022)
ACL