Fecha de publicación

2025-10-06T15:37:57Z

2025-10-06T15:37:57Z

2025



Resumen

Treball de fi de màster en Lingüística Teòrica i Aplicada


Directors: : Dra. Dra. Iria de Dios Flores i Dr. Corentin Kervadec


This study investigates how transformer-based large language models (LLMs) resolve subject–verb agreement across syntactic structures of varying complexity and distractor conditions. We hypothesized that LLMs would succeed in simple configurations but struggle under deeper embedding and number mismatches. Using a controlled dataset adapted from psycholinguistic research, we analyze model behavior across six sentence structures, two attractor conditions (mismatch vs. no mismatch), and four lexical variants. Using the Pythia 6.9B model, we apply three evaluation metrics—accuracy, prediction depth, and the Tuned Lens interpretability method—to track how agreement resolution evolves across layers. Results confirm our hypothesis: the model performs reliably in simple structures but fails in deeply embedded object-relative clauses. Prediction depth shows early resolution in simple cases and delayed or failed resolution in complex ones. These findings clarify LLM limitations in syntactic processing and highlight the importance of using linguistically informed evaluation methods to better understand model behavior across structural configurations.

Tipo de documento

Trabajo fin de máster

Lengua

Inglés

Materias y palabras clave

Psicolingüística

Citación recomendada

Esta citación se ha generado automáticamente.

Derechos

Llicència CC Reconeixement-NoComercial-SenseObraDerivada 4.0 Internacional (CC BY-NC-ND 4.0)

https://creativecommons.org/licenses/by-nc-nd/4.0/

Este ítem aparece en la(s) siguiente(s) colección(ones)