Can RNNs trained on harder subject-verb agreement instances still perform well on easier ones?
- Hritik Bansal (Indian Institute of Technology Delhi)
- Gantavya Bhatt (Indian Institute of Technology Delhi)
- Sumeet Agarwal (Indian Institute of Technology Delhi)
Abstract
The main subject and the associated verb in English must agree in grammatical number as per the Subject-Verb Agreement (SVA) phenomenon. It has been found that the presence of a noun between the verb and the main subject, whose grammatical number is opposite to that of the main subject, can cause speakers to produce a verb that agrees with the intervening noun rather than the main noun; the former thus acts as an agreement attractor. Such attractors have also been shown to pose a challenge for RNN models without explicit hierarchical bias to perform well on SVA tasks. Previous work suggests that syntactic cues in the input can aid such models to choose hierarchical rules over linear rules for number agreement. In this work, we investigate the effects of the choice of training data, training algorithm, and architecture on hierarchical generalization. We observe that the models under consideration fail to perform well on sentences with no agreement attractor when trained solely on natural sentences with at least one attractor. Even in the presence of this biased training set, implicit hierarchical bias in the architecture (as in the Ordered Neurons LSTM) is not enough to capture syntax-sensitive dependencies. These results suggest that current RNNs do not capture the underlying hierarchical rules of natural language, but rather use shallower heuristics for their predictions.
Keywords: subject-verb agreement, RNN models, generalisation
How to Cite:
Bansal, H., Bhatt, G. & Agarwal, S., (2021) “Can RNNs trained on harder subject-verb agreement instances still perform well on easier ones?”, Society for Computation in Linguistics 4(1), 366-373. doi: https://doi.org/10.7275/5bnr-wc78
Downloads:
Download PDF