Abstract
Speech translation systems usually follow a pipeline approach, using word lattices as an intermediate representation. However, previous work assume access to the original transcriptions used to train the ASR system, which can limit applicability in real scenarios. In this work we propose an approach for speech translation through lattice transformations and neural models based on graph networks. Experimental results show that our approach reaches competitive performance without relying on transcriptions, while also being orders of magnitude faster than previous work.- Anthology ID:
- D19-5304
- Volume:
- Proceedings of the Thirteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-13)
- Month:
- November
- Year:
- 2019
- Address:
- Hong Kong
- Editors:
- Dmitry Ustalov, Swapna Somasundaran, Peter Jansen, Goran Glavaš, Martin Riedl, Mihai Surdeanu, Michalis Vazirgiannis
- Venue:
- TextGraphs
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 26–31
- Language:
- URL:
- https://aclanthology.org/D19-5304
- DOI:
- 10.18653/v1/D19-5304
- Cite (ACL):
- Daniel Beck, Trevor Cohn, and Gholamreza Haffari. 2019. Neural Speech Translation using Lattice Transformations and Graph Networks. In Proceedings of the Thirteenth Workshop on Graph-Based Methods for Natural Language Processing (TextGraphs-13), pages 26–31, Hong Kong. Association for Computational Linguistics.
- Cite (Informal):
- Neural Speech Translation using Lattice Transformations and Graph Networks (Beck et al., TextGraphs 2019)
- PDF:
- https://preview.aclanthology.org/add_acl24_videos/D19-5304.pdf