Universitat Politècnica de Catalunya. Departament de Ciències de la Computació
Universitat Politècnica de Catalunya. LQMC - Lingüística Quantitativa, Matemàtica i Computacional
2025
In recent years, the need for natural language interfaces to knowledge graphs has become increasingly important since they enable easy and efficient access to the information contained in them. In particular, property graphs (PGs) have seen increased adoption as a means of representing complex structured information. Despite their growing popularity in industry, PGs remain relatively underrepresented in semantic parsing research with a lack of resources for evaluation. To address this gap, we introduce ZOGRASCOPE, a benchmark designed specifically for PGs and queries written in Cypher. Our benchmark includes a diverse set of manually annotated queries of varying complexity and is organized into three partitions: iid, compositional and length. We complement this paper with a set of experiments that test the performance of different LLMs in a variety of learning settings.
This project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme under grant agreement No 853459. The authors gratefully acknowledge the computer resources at ARTEMISA, funded by the European Union ERDF and Comunitat Valenciana as well as the technical support provided by the Instituto de Física Corpuscular, IFIC (CSIC-UV). This research is supported by a recognition 2021SGR-Cat (01266 LQMC) from AGAUR (Generalitat de Catalunya).
Peer Reviewed
Postprint (published version)
Conference report
Anglès
Àrees temàtiques de la UPC::Informàtica::Intel·ligència artificial::Llenguatge natural; Natural language; Property graphs (PGs); ZOGRASCOPE
Association for Computational Linguistics
https://aclanthology.org/2025.findings-emnlp.227/
info:eu-repo/grantAgreement/EC/H2020/853459/EU/Interactive Machine Learning for Compositional Models of Natural Language/INTERACT
http://creativecommons.org/licenses/by/4.0/
Open Access
E-prints [72263]