2024
pdf
abs
Comparing a BERT Classifier and a GPT classifier for Detecting Connective Language Across Multiple Social Media
Josephine Lukito
|
Bin Chen
|
Gina M. Masullo
|
Natalie Jomini Stroud
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
This study presents an approach for detecting connective language—defined as language that facilitates engagement, understanding, and conversation—from social media discussions. We developed and evaluated two types of classifiers: BERT and GPT-3.5 turbo. Our results demonstrate that the BERT classifier significantly outperforms GPT-3.5 turbo in detecting connective language. Furthermore, our analysis confirms that connective language is distinct from related concepts measuring discourse qualities, such as politeness and toxicity. We also explore the potential of BERT-based classifiers for platform-agnostic tools. This research advances our understanding of the linguistic dimensions of online communication and proposes practical tools for detecting connective language across diverse digital environments.
pdf
abs
Natural Evolution-based Dual-Level Aggregation for Temporal Knowledge Graph Reasoning
Bin Chen
|
Chunjing Xiao
|
Fan Zhou
Findings of the Association for Computational Linguistics: EMNLP 2024
Temporal knowledge graph (TKG) reasoning aims to predict missing facts based on a given history. Most of the existing methods unifiedly model the evolution process of different events and ignore their inherent asynchronous characteristics, resulting in suboptimal performance. To tackle this challenge, we propose a Natural Evolution-based Dual-level Aggregation framework (NEDA) for TKG reasoning. Specifically, we design a natural division strategy to group TKGs into different patches according to the occurrence of a given target entity. Then, we present a dual-level aggregation scheme to extract local representations from information within patches and then aggregate these representations with adaptive weights as the final entity representations. By assigning varying weights to different patches, this aggregation scheme can incorporate the asynchronous characteristics of event evolution for representation computation, thus enhancing prediction performance. Extensive experiments demonstrate the significant improvement of our proposed model.
pdf
abs
Decoding at the Speed of Thought: Harnessing Parallel Decoding of Lexical Units for LLMs
Chenxi Sun
|
Hongzhi Zhang
|
Zijia Lin
|
Jingyuan Zhang
|
Fuzheng Zhang
|
Zhongyuan Wang
|
Bin Chen
|
Chengru Song
|
Di Zhang
|
Kun Gai
|
Deyi Xiong
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Large language models have demonstrated exceptional capability in natural language understanding and generation. However, their generation speed is limited by the inherently sequential nature of their decoding process, posing challenges for real-time applications. This paper introduces Lexical Unit Decoding (LUD), a novel decoding methodology implemented in a data-driven manner, accelerating the decoding process without sacrificing output quality. The core of our approach is the observation that a pre-trained language model can confidently predict multiple contiguous tokens, forming the basis for a lexical unit, in which these contiguous tokens could be decoded in parallel. Extensive experiments validate that our method substantially reduces decoding time while maintaining generation quality, i.e., 33% speed up on natural language generation with no quality loss, and 30% speed up on code generation with a negligible quality loss of 3%. Distinctively, LUD requires no auxiliary models and does not require changes to existing architectures. It can also be integrated with other decoding acceleration methods, thus achieving an even more pronounced inference efficiency boost. We posit that the foundational principles of LUD could define a new decoding paradigm for future language models, enhancing their applicability for a broader spectrum of applications. All codes are be publicly available at https://github.com/tjunlp-lab/Lexical-Unit-Decoding-LUD-.
pdf
abs
Mitigating Linguistic Artifacts in Emotion Recognition for Conversations from TV Scripts to Daily Conversations
Donovan Ong
|
Shuo Sun
|
Jian Su
|
Bin Chen
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Emotion Recognition in Conversations (ERC) is a well-studied task with numerous potential real-world applications. However, existing ERC models trained on the MELD dataset derived from TV series, struggle when applied to daily conversation datasets. A closer examination of the datasets unveils the prevalence of linguistic artifacts such as repetitions and interjections in TV scripts, which ERC models may exploit when making predictions. To address this issue, we explore two techniques aimed at reducing the reliance of ERC models on these artifacts: 1) using contrastive learning to prioritize emotional features over dataset-specific linguistic style and 2) refining emotion predictions with pseudo-emotion intensity score. Our experiment results show that reducing reliance on the linguistic style found in TV transcripts could enhance model’s robustness and accuracy in diverse conversational contexts.
2023
pdf
abs
An Exploratory Study on Model Compression for Text-to-SQL
Shuo Sun
|
Yuze Gao
|
Yuchen Zhang
|
Jian Su
|
Bin Chen
|
Yingzhan Lin
|
Shuqi Sun
Findings of the Association for Computational Linguistics: ACL 2023
Text-to-SQL translates user queries into SQL statements that can retrieve relevant answers from relational databases. Recent approaches to Text-to-SQL rely on pre-trained language models that are computationally expensive and technically challenging to deploy in real-world applications that require real-time or on-device processing capabilities. In this paper, we perform a focused study on the feasibility of applying recent model compression techniques to sketch-based and sequence-to-sequence Text-to-SQL models. Our results reveal that sketch-based Text-to-SQL models generally have higher inference efficiency and respond better to model compression than sequence-to-sequence models, making them ideal for real-world deployments, especially in use cases with simple SQL statements.
pdf
abs
Battle of the Large Language Models: Dolly vs LLaMA vs Vicuna vs Guanaco vs Bard vs ChatGPT - A Text-to-SQL Parsing Comparison
Shuo Sun
|
Yuchen Zhang
|
Jiahuan Yan
|
Yuze Gao
|
Donovan Ong
|
Bin Chen
|
Jian Su
Findings of the Association for Computational Linguistics: EMNLP 2023
The success of ChatGPT has ignited an AI race, with researchers striving to develop new large language models (LLMs) that can match or surpass the language understanding and generation abilities of commercial ones. In recent times, a number of models have emerged, claiming performance near that of GPT-3.5 or GPT-4 through various instruction-tuning methods. As practitioners of Text-to-SQL parsing, we are grateful for their valuable contributions to open-source research. However, it is important to approach these claims with a sense of scrutiny and ascertain the actual effectiveness of these models. Therefore, we pit six popular large language models against each other, systematically evaluating their Text-to-SQL parsing capability on nine benchmark datasets with five different prompting strategies, covering both zero-shot and few-shot scenarios. Regrettably, the open-sourced models fell significantly short of the performance achieved by closed-source models like GPT-3.5, highlighting the need for further work to bridge the performance gap between these models.
2015
pdf
Improving Twitter Named Entity Recognition using Word Representations
Zhiqiang Toh
|
Bin Chen
|
Jian Su
Proceedings of the Workshop on Noisy User-generated Text
2013
pdf
bib
Exploiting Discourse Analysis for Article-Wide Temporal Classification
Jun-Ping Ng
|
Min-Yen Kan
|
Ziheng Lin
|
Wei Feng
|
Bin Chen
|
Jian Su
|
Chew-Lim Tan
Proceedings of the 2013 Conference on Empirical Methods in Natural Language Processing
2011
pdf
A Unified Event Coreference Resolution by Integrating Multiple Resolvers
Bin Chen
|
Jian Su
|
Sinno Jialin Pan
|
Chew Lim Tan
Proceedings of 5th International Joint Conference on Natural Language Processing
2010
pdf
A Twin-Candidate Based Approach for Event Pronoun Resolution using Composite Kernel
Bin Chen
|
Jian Su
|
Chew Lim Tan
Proceedings of the 23rd International Conference on Computational Linguistics (Coling 2010)
pdf
Resolving Event Noun Phrases to Their Verbal Mentions
Bin Chen
|
Jian Su
|
Chew Lim Tan
Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing
2008
pdf
Other-Anaphora Resolution in Biomedical Texts with Automatically Mined Patterns
Bin Chen
|
Xiaofeng Yang
|
Jian Su
|
Chew Lim Tan
Proceedings of the 22nd International Conference on Computational Linguistics (Coling 2008)