Sequential Neural Networks as Automata

William Merrill


Abstract
This work attempts to explain the types of computation that neural networks can perform by relating them to automata. We first define what it means for a real-time network with bounded precision to accept a language. A measure of network memory follows from this definition. We then characterize the classes of languages acceptable by various recurrent networks, attention, and convolutional networks. We find that LSTMs function like counter machines and relate convolutional networks to the subregular hierarchy. Overall, this work attempts to increase our understanding and ability to interpret neural networks through the lens of theory. These theoretical insights help explain neural computation, as well as the relationship between neural networks and natural language grammar.
Anthology ID:
W19-3901
Volume:
Proceedings of the Workshop on Deep Learning and Formal Languages: Building Bridges
Month:
August
Year:
2019
Address:
Florence
Editors:
Jason Eisner, Matthias Gallé, Jeffrey Heinz, Ariadna Quattoni, Guillaume Rabusseau
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1–13
Language:
URL:
https://aclanthology.org/W19-3901
DOI:
10.18653/v1/W19-3901
Bibkey:
Cite (ACL):
William Merrill. 2019. Sequential Neural Networks as Automata. In Proceedings of the Workshop on Deep Learning and Formal Languages: Building Bridges, pages 1–13, Florence. Association for Computational Linguistics.
Cite (Informal):
Sequential Neural Networks as Automata (Merrill, ACL 2019)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-acl-2023-videos/W19-3901.pdf