@inproceedings{razumovskaia-etal-2022-data,
title = "Data Augmentation and Learned Layer Aggregation for Improved Multilingual Language Understanding in Dialogue",
author = "Razumovskaia, Evgeniia and
Vuli{\'c}, Ivan and
Korhonen, Anna",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.160/",
doi = "10.18653/v1/2022.findings-acl.160",
pages = "2017--2033",
abstract = "Scaling dialogue systems to a multitude of domains, tasks and languages relies on costly and time-consuming data annotation for different domain-task-language configurations. The annotation efforts might be substantially reduced by the methods that generalise well in zero- and few-shot scenarios, and also effectively leverage external unannotated data sources (e.g., Web-scale corpora). We propose two methods to this aim, offering improved dialogue natural language understanding (NLU) across multiple languages: 1) Multi-SentAugment, and 2) LayerAgg. Multi-SentAugment is a self-training method which augments available (typically few-shot) training data with similar (automatically labelled) in-domain sentences from large monolingual Web-scale corpora. LayerAgg learns to select and combine useful semantic information scattered across different layers of a Transformer model (e.g., mBERT); it is especially suited for zero-shot scenarios as semantically richer representations should strengthen the model{'}s cross-lingual capabilities. Applying the two methods with state-of-the-art NLU models obtains consistent improvements across two standard multilingual NLU datasets covering 16 diverse languages. The gains are observed in zero-shot, few-shot, and even in full-data scenarios. The results also suggest that the two methods achieve a synergistic effect: the best overall performance in few-shot setups is attained when the methods are used together."
}
Markdown (Informal)
[Data Augmentation and Learned Layer Aggregation for Improved Multilingual Language Understanding in Dialogue](https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.160/) (Razumovskaia et al., Findings 2022)
ACL