TY - JOUR
T1 - Decoding BERT’s Internal Processing of Garden-Path Structures through Attention Maps*
AU - Lee, Jonghyun
AU - Shin, Jeong Ah
N1 - Publisher Copyright:
© 2023 KASELL. All rights reserved.
PY - 2023
Y1 - 2023
N2 - Recent advancements in deep learning neural models, such as BERT, have demonstrated remarkable performance in natural language processing tasks, yet understanding their internal processing remains a challenge. This study employs the method of examining attention maps to uncover the internal processing of BERT, specifically when dealing with garden-path sentences. The analysis focuses on BERT's utilization of linguistic cues, such as transitivity, plausibility, and the presence of a comma, and evaluates its capacity for reanalyzing misinterpretations. The results revealed that BERT exhibits human-like syntactic processing by attending to the presence of a comma, showing sensitivity to transitivity, and reanalyzing misinterpretations, despite initially lacking sensitivity to plausibility. By concentrating on attention maps, the present study provides valuable insights into the inner workings of BERT and contributes to a deeper understanding of how advanced neural language models acquire and process complex linguistic structures.
AB - Recent advancements in deep learning neural models, such as BERT, have demonstrated remarkable performance in natural language processing tasks, yet understanding their internal processing remains a challenge. This study employs the method of examining attention maps to uncover the internal processing of BERT, specifically when dealing with garden-path sentences. The analysis focuses on BERT's utilization of linguistic cues, such as transitivity, plausibility, and the presence of a comma, and evaluates its capacity for reanalyzing misinterpretations. The results revealed that BERT exhibits human-like syntactic processing by attending to the presence of a comma, showing sensitivity to transitivity, and reanalyzing misinterpretations, despite initially lacking sensitivity to plausibility. By concentrating on attention maps, the present study provides valuable insights into the inner workings of BERT and contributes to a deeper understanding of how advanced neural language models acquire and process complex linguistic structures.
KW - attention map
KW - garden-path structure
KW - Natural Language Processing
KW - Psycholinguistics
KW - Transformers
UR - http://www.scopus.com/inward/record.url?scp=85166292757&partnerID=8YFLogxK
U2 - 10.15738/kjell.23..202306.461
DO - 10.15738/kjell.23..202306.461
M3 - Article
AN - SCOPUS:85166292757
SN - 1598-1398
VL - 23
SP - 461
EP - 481
JO - Korean Journal of English Language and Linguistics
JF - Korean Journal of English Language and Linguistics
ER -