Neural Networks as Cognitive Models of the Processing of Syntactic Constraints
- PMID: 38746852
- PMCID: PMC11093404
- DOI: 10.1162/opmi_a_00137
Neural Networks as Cognitive Models of the Processing of Syntactic Constraints
Abstract
Languages are governed by syntactic constraints-structural rules that determine which sentences are grammatical in the language. In English, one such constraint is subject-verb agreement, which dictates that the number of a verb must match the number of its corresponding subject: "the dogs run", but "the dog runs". While this constraint appears to be simple, in practice speakers make agreement errors, particularly when a noun phrase near the verb differs in number from the subject (for example, a speaker might produce the ungrammatical sentence "the key to the cabinets are rusty"). This phenomenon, referred to as agreement attraction, is sensitive to a wide range of properties of the sentence; no single existing model is able to generate predictions for the wide variety of materials studied in the human experimental literature. We explore the viability of neural network language models-broad-coverage systems trained to predict the next word in a corpus-as a framework for addressing this limitation. We analyze the agreement errors made by Long Short-Term Memory (LSTM) networks and compare them to those of humans. The models successfully simulate certain results, such as the so-called number asymmetry and the difference between attraction strength in grammatical and ungrammatical sentences, but failed to simulate others, such as the effect of syntactic distance or notional (conceptual) number. We further evaluate networks trained with explicit syntactic supervision, and find that this form of supervision does not always lead to more human-like syntactic behavior. Finally, we show that the corpus used to train a network significantly affects the pattern of agreement errors produced by the network, and discuss the strengths and limitations of neural networks as a tool for understanding human syntactic processing.
Keywords: agreement attraction; computational modeling; neural networks; psycholinguistics; syntactic processing.
© 2024 Massachusetts Institute of Technology.
Conflict of interest statement
Competing Interests: The authors declare no conflict of interests.
Figures
References
-
- Badecker, W., & Kuminiak, F. (2007). Morphology, agreement and working memory retrieval in sentence production: Evidence from gender and case in Slovak. Journal of Memory and Language, 56(1), 65–85. 10.1016/j.jml.200608004 - DOI
-
- Bangalore, S., & Joshi, A. K. (1999). Supertagging: An approach to almost parsing. Computational Linguistics, 25(2), 237–265.
-
- Bender, E. M., & Koller, A. (2020). Climbing towards NLU: On meaning, form, and understanding in the age of data. In Jurafsky D., Chai J., Schluter N., & Tetreault J. (Eds.), Proceedings of 58th Annual Meeting of the Association for Computational Linguistics (pp. 5185–5198). Association for Computational Linguistics. 10.18653/v1/2020.acl-main.463 - DOI
-
- Bhatt, G., Bansal, H., Singh, R., & Agarwal, S. (2020). How much complexity does an RNN architecture need to learn syntax-sensitive dependencies? In Rijhwani S., Liu J., Wang Y., & Dror R. (Eds.), Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics: Student Research Workshop (pp. 244–254). Association for Computational Linguistics. 10.18653/v1/2020.acl-srw.33 - DOI
-
- Bock, K., & Cutting, J. C. (1992). Regulating mental energy: Performance units in language production. Journal of Memory and Language, 31(1), 99–127. 10.1016/0749-596X(92)90007-K - DOI
LinkOut - more resources
Full Text Sources