LLM4VKG: Leveraging Large Language Models for Virtual Knowledge Graph Construction

LLM4VKG: Leveraging Large Language Models for Virtual Knowledge Graph Construction

Guohui Xiao, Lin Ren, Guilin Qi, Haohan Xue, Marco Di Panfilo, Davide Lanti

Proceedings of the Thirty-Fourth International Joint Conference on Artificial Intelligence
Main Track. Pages 4715-4723. https://doi.org/10.24963/ijcai.2025/525

Virtual Knowledge Graphs (VKGs) provide an effective solution for data integration but typically require significant expertise for their construction. This process, involving ontology development, schema analysis, and mapping creation, is often hindered by naming ambiguities and matching issues, which traditional rule-based methods struggle to address. Large language models (LLMs), with their ability to process and generate contextually relevant text, offer a potential solution. In this work, we introduce LLM4VKG, a novel framework that leverages LLMs to automatize VKG construction. Experimental evaluation on the RODI benchmark demonstrates that LLM4VKG surpasses state-of-the-art methods, achieving an average F1-score improvement of +17% and a peak gain of +39%. Moreover, LLM4VKG proves robust against incomplete ontologies and can handle complex mappings where current methods fail.
Keywords:
Knowledge Representation and Reasoning: KRR: Semantic Web
Data Mining: DM: Knowledge graphs and knowledge base completion
Knowledge Representation and Reasoning: KRR: Applications
Natural Language Processing: NLP: Language models