dc.contributor
Universitat Ramon Llull. Esade
dc.contributor.author
Bermejo, Vicente J.
dc.contributor.author
Gago, Andrés
dc.contributor.author
Gálvez, Ramiro H.
dc.contributor.author
Harari, Nicolás
dc.date.accessioned
2026-03-06T20:00:12Z
dc.date.available
2026-03-06T20:00:12Z
dc.date.issued
2025-11-17
dc.identifier.issn
2045-2322
dc.identifier.uri
https://hdl.handle.net/20.500.14342/6013
dc.description.abstract
This paper evaluates the effectiveness of large language models (LLMs) in extracting complex information from text data. Using a corpus of Spanish news articles, we compare how accurately various LLMs and outsourced human coders reproduce expert annotations on five natural language processing tasks, ranging from named entity recognition to identifying nuanced political criticism in news articles. We find that LLMs consistently outperform outsourced human coders, particularly in tasks requiring deep contextual understanding. These findings suggest that current LLM technology offers researchers without programming expertise a cost-effective alternative for sophisticated text analysis.
dc.publisher
Springer Nature
dc.relation.ispartof
Scientific Reports, Vol. 15, 40122
dc.rights
Attribution-NonCommercial-NoDerivatives 4.0 International
dc.rights.uri
http://creativecommons.org/licenses/by-nc-nd/4.0/
dc.subject
Natural Language Processing
dc.title
LLMs outperform outsourced human coders on complex textual analysis
dc.type
info:eu-repo/semantics/article
dc.description.version
info:eu-repo/semantics/publishedVersion
dc.identifier.doi
https://doi.org/10.1038/s41598-025-23798-y
dc.rights.accessLevel
info:eu-repo/semantics/openAccess