Kim Cheng Sheang


2021

pdf bib
Controllable Sentence Simplification with a Unified Text-to-Text Transfer Transformer
Kim Cheng Sheang | Horacio Saggion
Proceedings of the 14th International Conference on Natural Language Generation

Recently, a large pre-trained language model called T5 (A Unified Text-to-Text Transfer Transformer) has achieved state-of-the-art performance in many NLP tasks. However, no study has been found using this pre-trained model on Text Simplification. Therefore in this paper, we explore the use of T5 fine-tuning on Text Simplification combining with a controllable mechanism to regulate the system outputs that can help generate adapted text for different target audiences. Our experiments show that our model achieves remarkable results with gains of between +0.69 and +1.41 over the current state-of-the-art (BART+ACCESS). We argue that using a pre-trained model such as T5, trained on several tasks with large amounts of data, can help improve Text Simplification.

2019

pdf bib
Multilingual Complex Word Identification: Convolutional Neural Networks with Morphological and Linguistic Features
Kim Cheng Sheang
Proceedings of the Student Research Workshop Associated with RANLP 2019

The paper is about our experiments with Complex Word Identification system using deep learning approach with word embeddings and engineered features.
Search
Co-authors
Venues