Skip to main navigation Skip to search Skip to main content

Probing Good-Enough Processing in Large Language Models with a Paraphrasing Task

Research output: Contribution to journalArticlepeer-review

Abstract

This study investigates whether large language models (LLMs) exhibit human-like ‘good-enough’ processing patterns in syntactic comprehension or demonstrate mechanical accuracy. Previous research using forced-choice question-answering paradigms revealed that LLMs display incomplete syntactic reanalysis similar to humans when processing garden-path sentences. However, concerns arose that these patterns might reflect methodological artifacts rather than genuine processing characteristics, as direct questioning could bias models toward initial misinterpretations. To address this limitation, we employed a paraphrasing task that requires comprehensive sentence reformulation rather than binary responses, following Patson et al. (2009). We tested GPT-3.5 and GPT-4 on 24 garden-path sentences containing Optionally Transitive (OT) and Reflexive Absolute Transitive (RAT) verbs. Results demonstrate that good-enough processing patterns persist across both paradigms, with LLMs continuing to exhibit partial reanalysis in garden-path conditions even when generating full paraphrases. This confirms that previously observed error patterns represent genuine syntactic processing characteristics rather than experimental artifacts. Notably, GPT-4 showed improved performance in the paraphrasing task compared to forced-choice experiments, suggesting task-dependent variation in processing depth. Both models exhibited human-like incomplete processing despite their substantial computational resources, indicating that their pattern-matching mechanisms favor processing shortcuts over complete syntactic interpretation. These findings reveal that LLMs demonstrate good-enough processing similar to humans, with performance varying systematically across task formats.

Original languageEnglish
Pages (from-to)127-141
Number of pages15
JournalKorean Journal of English Language and Linguistics
Volume26
DOIs
StatePublished - 2026

Keywords

  • ChatGPT
  • garden-path sentences
  • good-enough processing
  • large language models
  • paraphrasing task
  • syntactic processing

Fingerprint

Dive into the research topics of 'Probing Good-Enough Processing in Large Language Models with a Paraphrasing Task'. Together they form a unique fingerprint.

Cite this