Sharanya Chakravarthy
2020
Domino at FinCausal 2020, Task 1 and 2: Causal Extraction System
Sharanya Chakravarthy
|
Tushar Kanakagiri
|
Karthik Radhakrishnan
|
Anjana Umapathy
Proceedings of the 1st Joint Workshop on Financial Narrative Processing and MultiLing Financial Summarisation
Automatic identification of cause-effect relationships from data is a challenging but important problem in artificial intelligence. Identifying semantic relationships has become increasingly important for multiple downstream applications like Question Answering, Information Retrieval and Event Prediction. In this work, we tackle the problem of causal relationship extraction from financial news using the FinCausal 2020 dataset. We tackle two tasks - 1) Detecting the presence of causal relationships and 2) Extracting segments corresponding to cause and effect from news snippets. We propose Transformer based sequence and token classification models with post-processing rules which achieve an F1 score of 96.12 and 79.60 on Tasks 1 and 2 respectively.
Detecting Entailment in Code-Mixed Hindi-English Conversations
Sharanya Chakravarthy
|
Anjana Umapathy
|
Alan W Black
Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020)
The presence of large-scale corpora for Natural Language Inference (NLI) has spurred deep learning research in this area, though much of this research has focused solely on monolingual data. Code-mixing is the intertwined usage of multiple languages, and is commonly seen in informal conversations among polyglots. Given the rising importance of dialogue agents, it is imperative that they understand code-mixing, but the scarcity of code-mixed Natural Language Understanding (NLU) datasets has precluded research in this area. The dataset by Khanuja et. al. for detecting conversational entailment in code-mixed Hindi-English text is the first of its kind. We investigate the effectiveness of language modeling, data augmentation, translation, and architectural approaches to address the code-mixed, conversational, and low-resource aspects of this dataset. We obtain an 8.09% increase in test set accuracy over the current state of the art.
“A Little Birdie Told Me ... ” - Inductive Biases for Rumour Stance Detection on Social Media
Karthik Radhakrishnan
|
Tushar Kanakagiri
|
Sharanya Chakravarthy
|
Vidhisha Balachandran
Proceedings of the Sixth Workshop on Noisy User-generated Text (W-NUT 2020)
The rise in the usage of social media has placed it in a central position for news dissemination and consumption. This greatly increases the potential for proliferation of rumours and misinformation. In an effort to mitigate the spread of rumours, we tackle the related task of identifying the stance (Support, Deny, Query, Comment) of a social media post. Unlike previous works, we impose inductive biases that capture platform specific user behavior. These biases, coupled with social media fine-tuning of BERT allow for better language understanding, thus yielding an F1 score of 58.7 on the SemEval 2019 task on rumour stance detection.