TY - JOUR
T1 - (AL)BERT Down the Garden Path
T2 - Psycholinguistic Experiments for Pre-trained Language Models
AU - Lee, Jonghyun
AU - Shin, Jeong Ah
AU - Park, Myung Kwan
N1 - Publisher Copyright:
© 2022 KASELL All rights reserved.
PY - 2022
Y1 - 2022
N2 - This study compared the syntactic capabilities of several neural language models (LMs) including Transformers (BERT / ALBERT) and LSTM and investigated whether they exhibit human-like syntactic representations through a targeted evaluation approach, a method to evaluate the syntactic processing ability of LMs using sentences designed for psycholinguistic experiments. By employing garden-path structures with several linguistic manipulations, whether LMs detect temporary ungrammaticality and use a linguistic cue such as plausibility, transitivity, and morphology is assessed. The results showed that both Transformers and LSTM exploited several linguistic cues for incremental syntactic processing, comparable to human syntactic processing. They differed, however, in terms of whether and how they use each linguistic cue. Overall, Transformers had a more human-like syntactic representation than LSTM, given their higher sensitivity to plausibility and ability to retain information from previous words. Meanwhile, the number of parameters does not seem to undermine the performance of LMs, contrary to what was predicted in previous studies. Through these findings, this research sought to contribute to a greater understanding of the syntactic processing of neural language models as well as human language processing.
AB - This study compared the syntactic capabilities of several neural language models (LMs) including Transformers (BERT / ALBERT) and LSTM and investigated whether they exhibit human-like syntactic representations through a targeted evaluation approach, a method to evaluate the syntactic processing ability of LMs using sentences designed for psycholinguistic experiments. By employing garden-path structures with several linguistic manipulations, whether LMs detect temporary ungrammaticality and use a linguistic cue such as plausibility, transitivity, and morphology is assessed. The results showed that both Transformers and LSTM exploited several linguistic cues for incremental syntactic processing, comparable to human syntactic processing. They differed, however, in terms of whether and how they use each linguistic cue. Overall, Transformers had a more human-like syntactic representation than LSTM, given their higher sensitivity to plausibility and ability to retain information from previous words. Meanwhile, the number of parameters does not seem to undermine the performance of LMs, contrary to what was predicted in previous studies. Through these findings, this research sought to contribute to a greater understanding of the syntactic processing of neural language models as well as human language processing.
KW - garden-path structure
KW - natural language processing
KW - psycholinguistics
KW - targeted evaluation approach
KW - transformers
UR - http://www.scopus.com/inward/record.url?scp=85139740453&partnerID=8YFLogxK
U2 - 10.15738/kjell.22..202210.1033
DO - 10.15738/kjell.22..202210.1033
M3 - Article
AN - SCOPUS:85139740453
SN - 1598-1398
VL - 22
SP - 1033
EP - 1050
JO - Korean Journal of English Language and Linguistics
JF - Korean Journal of English Language and Linguistics
ER -