Skip Nav Destination
Close Modal
Update search
NARROW
Format
Journal
Date
Availability
1-1 of 1
Samuel Barrett
Close
Follow your search
Access your saved searches in your account
Would you like to receive an alert when new items match your search?
Sort by
Journal Articles
Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
Open AccessPublisher: Journals Gateway
Transactions of the Association for Computational Linguistics (2022) 10: 1423–1439.
Published: 22 December 2022
FIGURES
| View all 5
Abstract
View articletitled, Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
View
PDF
for article titled, Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
We introduce Transformer Grammars (TGs), a novel class of Transformer language models that combine (i) the expressive power, scalability, and strong performance of Transformers and (ii) recursive syntactic compositions, which here are implemented through a special attention mask and deterministic transformation of the linearized tree. We find that TGs outperform various strong baselines on sentence-level language modeling perplexity, as well as on multiple syntax-sensitive language modeling evaluation metrics. Additionally, we find that the recursive syntactic composition bottleneck which represents each sentence as a single vector harms perplexity on document-level language modeling, providing evidence that a different kind of memory mechanism—one that is independent of composed syntactic representations—plays an important role in current successful models of long text.