Maria Kustikova
2019
Comparison of Diverse Decoding Methods from Conditional Language Models
Daphne Ippolito
|
Reno Kriz
|
João Sedoc
|
Maria Kustikova
|
Chris Callison-Burch
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics
While conditional language models have greatly improved in their ability to output high quality natural language, many NLP applications benefit from being able to generate a diverse set of candidate sequences. Diverse decoding strategies aim to, within a given-sized candidate list, cover as much of the space of high-quality outputs as possible, leading to improvements for tasks that rerank and combine candidate outputs. Standard decoding methods, such as beam search, optimize for generating high likelihood sequences rather than diverse ones, though recent work has focused on increasing diversity in these methods. In this work, we perform an extensive survey of decoding-time strategies for generating diverse outputs from a conditional language model. In addition, we present a novel method where we over-sample candidates, then use clustering to remove similar sequences, thus achieving high diversity without sacrificing quality.
2018
Simple Features for Strong Performance on Named Entity Recognition in Code-Switched Twitter Data
Devanshu Jain
|
Maria Kustikova
|
Mayank Darbari
|
Rishabh Gupta
|
Stephen Mayhew
Proceedings of the Third Workshop on Computational Approaches to Linguistic Code-Switching
In this work, we address the problem of Named Entity Recognition (NER) in code-switched tweets as a part of the Workshop on Computational Approaches to Linguistic Code-switching (CALCS) at ACL’18. Code-switching is the phenomenon where a speaker switches between two languages or variants of the same language within or across utterances, known as intra-sentential or inter-sentential code-switching, respectively. Processing such data is challenging using state of the art methods since such technology is generally geared towards processing monolingual text. In this paper we explored ways to use language identification and translation to recognize named entities in such data, however, utilizing simple features (sans multi-lingual features) with Conditional Random Field (CRF) classifier achieved the best results. Our experiments were mainly aimed at the (ENG-SPA) English-Spanish dataset but we submitted a language-independent version of our system to the (MSA-EGY) Arabic-Egyptian dataset as well and achieved good results.
Search
Co-authors
- Daphne Ippolito 1
- Reno Kriz 1
- João Sedoc 1
- Chris Callison-Burch 1
- Devanshu Jain 1
- show all...
Venues
- acl2