Skip to content

dice-group/KATRINA

Repository files navigation

UniQ-Gen

This repository contains the code for the paper UniQ-Gen: Unified Query Generation across Multiple Knowledge Graphs. The goal is to train unified query generation models across knowledge graphs. To this end, this repository contains code for training query generation models based on LLMs in various setups and with different input from entity linking systems.

Installation

  • Dowload the repository
  • Run pip install -r requirements

Files

All required files including trained models can be downloaded from our FTP server: https://files.dice-research.org/projects/UniQ-Gen/EKAW/

Training Query Generation models

Run the script train.py in the package Generator Model check parameters in the python file parameters.py for joint and single model training, for editing paths to the datasets and other configurations

for joint training the files for both traing datasets should be in the same folder.

Entity Linking

Wikidata

Freebase/GrailQA

Experiments

How to cite:

@InProceedings{10.1007/978-3-031-77792-9_11,
author="Vollmers, Daniel
and Srivastava, Nikit
and Zahera, Hamada M.
and Moussallem, Diego
and Ngomo, Axel-Cyrille Ngonga",
editor="Alam, Mehwish
and Rospocher, Marco
and van Erp, Marieke
and Hollink, Laura
and Gesese, Genet Asefa",
title="UniQ-Gen: Unified Query Generation Across Multiple Knowledge Graphs",
booktitle="Knowledge Engineering and Knowledge Management",
year="2025",
publisher="Springer Nature Switzerland",
address="Cham",
pages="174--189",
abstract="Generating SPARQL queries is crucial for extracting relevant information from diverse knowledge graphs. However, the structural and semantic differences among these graphs necessitate training or fine-tuning a tailored model for each one. In this paper, we propose UniQ-Gen, a unified query generation approach to generate SPARQL queries across various knowledge graphs. UniQ-Gen integrates entity recognition, disambiguation, and linking through a BERT-NER model and employs cross-encoder ranking to align questions with the Freebase ontology. We conducted several experiments on different benchmark datasets such as LC-QuAD 2.0, GrailQA, and QALD-10. The evaluation results demonstrate that our approach achieves performance equivalent to or better than models fine-tuned for individual knowledge graphs. This finding suggests that fine-tuning a unified model on a heterogeneous dataset of SPARQL queries across different knowledge graphs eliminates the need for separate models for each graph, thereby reducing resource requirements.",
isbn="978-3-031-77792-9"
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages