Dive into Deep Learning for Natural Language Processing
Haibin Lin, Xingjian Shi, Leonard Lausen, Aston Zhang, He He, Sheng Zha, Alexander Smola
Abstract
Deep learning has become the dominant approach to NLP problems, especially when applied on large scale corpora. Recent progress on unsupervised pre-training techniques such as BERT, ELMo, GPT-2, and language modeling in general, when applied on large corpora, is shown to be effective in improving a wide variety of downstream tasks. These techniques push the limits of available hardware, requiring specialized frameworks optimized for GPU, ASIC, and distributed cloud-based training.A few complexities pose challenges to scale these models and algorithms effectively. Compared to other areas where deep learning is applied, these NLP models contain a variety of moving parts: text normalization and tokenization, word representation at subword-level and word-level, variable-length models such as RNN and attention, and sequential decoder based on beam search, among others.In this hands-on tutorial, we take a closer look at the challenges from these complexities and see how with proper tooling with Apache MXNet and GluonNLP, we can overcome these challenges and achieve state-of-the-art results for real-world problems. GluonNLP is a powerful new toolkit that combines MXNet’s speed, the flexibility of Gluon, and an extensive new library automating the most laborious aspects of deep learning for NLP.- Anthology ID:
- D19-2001
- Volume:
- Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): Tutorial Abstracts
- Month:
- November
- Year:
- 2019
- Address:
- Hong Kong, China
- Editors:
- Timothy Baldwin, Marine Carpuat
- Venues:
- EMNLP | IJCNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- Language:
- URL:
- https://aclanthology.org/D19-2001
- DOI:
- Cite (ACL):
- Haibin Lin, Xingjian Shi, Leonard Lausen, Aston Zhang, He He, Sheng Zha, and Alexander Smola. 2019. Dive into Deep Learning for Natural Language Processing. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP): Tutorial Abstracts, Hong Kong, China. Association for Computational Linguistics.
- Cite (Informal):
- Dive into Deep Learning for Natural Language Processing (Lin et al., EMNLP-IJCNLP 2019)