Nghia Ngo Trung
Also published as: Nghia Ngo Trung
2022
Unsupervised Domain Adaptation for Text Classification via Meta Self-Paced Learning
Nghia Ngo Trung
|
Linh Ngo Van
|
Thien Huu Nguyen
Proceedings of the 29th International Conference on Computational Linguistics
A shift in data distribution can have a significant impact on performance of a text classification model. Recent methods addressing unsupervised domain adaptation for textual tasks typically extracted domain-invariant representations through balancing between multiple objectives to align feature spaces between source and target domains. While effective, these methods induce various new domain-sensitive hyperparameters, thus are impractical as large-scale language models are drastically growing bigger to achieve optimal performance. To this end, we propose to leverage meta-learning framework to train a neural network-based self-paced learning procedure in an end-to-end manner. Our method, called Meta Self-Paced Domain Adaption (MSP-DA), follows a novel but intuitive domain-shift variation of cluster assumption to derive the meta train-test dataset split based on the self-pacing difficulties of source domain’s examples. As a result, MSP-DA effectively leverages self-training and self-tuning domain-specific hyperparameters simultaneously throughout the learning process. Extensive experiments demonstrate our framework substantially improves performance on target domains, surpassing state-of-the-art approaches. Detailed analyses validate our method and provide insight into how each domain affects the learned hyperparameters.
2021
Unsupervised Domain Adaptation for Event Detection using Domain-specific Adapters
Nghia Ngo Trung
|
Duy Phung
|
Thien Huu Nguyen
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
Modeling Document-Level Context for Event Detection via Important Context Selection
Amir Pouran Ben Veyseh
|
Minh Van Nguyen
|
Nghia Ngo Trung
|
Bonan Min
|
Thien Huu Nguyen
Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
The task of Event Detection (ED) in Information Extraction aims to recognize and classify trigger words of events in text. The recent progress has featured advanced transformer-based language models (e.g., BERT) as a critical component in state-of-the-art models for ED. However, the length limit for input texts is a barrier for such ED models as they cannot encode long-range document-level context that has been shown to be beneficial for ED. To address this issue, we propose a novel method to model document-level context for ED that dynamically selects relevant sentences in the document for the event prediction of the target sentence. The target sentence will be then augmented with the selected sentences and consumed entirely by transformer-based language models for improved representation learning for ED. To this end, the REINFORCE algorithm is employed to train the relevant sentence selection for ED. Several information types are then introduced to form the reward function for the training process, including ED performance, sentence similarity, and discourse relations. Our extensive experiments on multiple benchmark datasets reveal the effectiveness of the proposed model, leading to new state-of-the-art performance.
Search
Co-authors
- Thien Huu Nguyen 3
- Linh Ngo Van 1
- Duy Phung 1
- Amir Pouran Ben Veyseh 1
- Minh Van Nguyen 1
- show all...