Publication Date
1-1-2022
Document Type
Conference Proceeding
Publication Title
Proceedings - International Conference on Computational Linguistics, COLING
Volume
29
Issue
1
First Page
6007
Last Page
6018
Abstract
Recombining known primitive concepts into larger novel combinations is a quintessentially human cognitive capability. Whether large neural models in NLP can acquire this ability while learning from data is an open question. In this paper, we investigate this problem from the perspective of formal languages. We use deterministic finite-state transducers to make an unbounded number of datasets with controllable properties governing compositionality. By randomly sampling over many transducers, we explore which of their properties contribute to learnability of a compositional relation by a neural network. We find that the models either learn the relations completely or not at all. The key is transition coverage, setting a soft learnability limit at 400 examples per transition.
Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 License.
Department
Linguistics and Language Development
Recommended Citation
Josef Valvoda, Naomi Saphra, Jonathan Rawski, Adina Williams, and Ryan Cotterell. "Benchmarking Compositionality with Formal Languages" Proceedings - International Conference on Computational Linguistics, COLING (2022): 6007-6018.