Abstract
Spoken Language Understanding (SLU) models in industry applications are usually trained offline on historic data, but have to perform well on incoming user requests after deployment. Since the application data is not available at training time, this is formally similar to the domain generalization problem, where domains correspond to different temporal segments of the data, and the goal is to build a model that performs well on unseen domains, e.g., upcoming data. In this paper, we explore different strategies for achieving good temporal generalization, including instance weighting, temporal fine-tuning, learning temporal features and building a temporally-invariant model. Our results on data of large-scale SLU systems show that temporal information can be leveraged to improve temporal generalization for SLU models.- Anthology ID:
- 2022.naacl-industry.5
- Volume:
- Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Track
- Month:
- July
- Year:
- 2022
- Address:
- Hybrid: Seattle, Washington + Online
- Editors:
- Anastassia Loukina, Rashmi Gangadharaiah, Bonan Min
- Venue:
- NAACL
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 37–44
- Language:
- URL:
- https://aclanthology.org/2022.naacl-industry.5
- DOI:
- 10.18653/v1/2022.naacl-industry.5
- Cite (ACL):
- Judith Gaspers, Anoop Kumar, Greg Ver Steeg, and Aram Galstyan. 2022. Temporal Generalization for Spoken Language Understanding. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Industry Track, pages 37–44, Hybrid: Seattle, Washington + Online. Association for Computational Linguistics.
- Cite (Informal):
- Temporal Generalization for Spoken Language Understanding (Gaspers et al., NAACL 2022)
- PDF:
- https://preview.aclanthology.org/nschneid-patch-3/2022.naacl-industry.5.pdf