QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries
Answering logical questions with a knowledge graph has been a critical research focus because this needs to reason and synthesize information. Previous studies have mainly dealt with logical operations using graph embedding techniques, such as conjunctions, disjunctions, and negation. However, these...
Saved in:
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2025-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/10836698/ |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832583247968600064 |
---|---|
author | Truong H. V. Phan Phuc do |
author_facet | Truong H. V. Phan Phuc do |
author_sort | Truong H. V. Phan |
collection | DOAJ |
description | Answering logical questions with a knowledge graph has been a critical research focus because this needs to reason and synthesize information. Previous studies have mainly dealt with logical operations using graph embedding techniques, such as conjunctions, disjunctions, and negation. However, these studies have neither effectively organized the data to retrieve multi-hop reasoning quickly nor combined text description to enhance logical operations’ semantics. Our study introduces a model called QUERY2BERT, which solves two of the above limitations. Specifically, QUERY2BERT first combined the node2vec and the BERT models to embed a knowledge graph with description information of every entity. Then, embedded nodes were indexed with a K-D tree structure. Finally, we used nearest neighbor search on K-D tree to retrieve neighbor-embedded nodes and implemented logical operations like projection, intersection, union, and negation to find answers to complex questions. We tested our model on three benchmark knowledge graph datasets and showed that QUERY2BERT significantly improved accuracy and speed compared to other state-of-the-art models. |
format | Article |
id | doaj-art-accf26e79ca74117aa99853c86c45518 |
institution | Kabale University |
issn | 2169-3536 |
language | English |
publishDate | 2025-01-01 |
publisher | IEEE |
record_format | Article |
series | IEEE Access |
spelling | doaj-art-accf26e79ca74117aa99853c86c455182025-01-29T00:01:15ZengIEEEIEEE Access2169-35362025-01-0113161031611910.1109/ACCESS.2025.352809710836698QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical QueriesTruong H. V. Phan0https://orcid.org/0000-0003-1773-6703Phuc do1https://orcid.org/0000-0001-6475-8716Department of Information Systems, University of Information Technology, Viet Nam National University Ho Chi Minh City, Ho Chi Minh City, VietnamDepartment of Information Systems, University of Information Technology, Viet Nam National University Ho Chi Minh City, Ho Chi Minh City, VietnamAnswering logical questions with a knowledge graph has been a critical research focus because this needs to reason and synthesize information. Previous studies have mainly dealt with logical operations using graph embedding techniques, such as conjunctions, disjunctions, and negation. However, these studies have neither effectively organized the data to retrieve multi-hop reasoning quickly nor combined text description to enhance logical operations’ semantics. Our study introduces a model called QUERY2BERT, which solves two of the above limitations. Specifically, QUERY2BERT first combined the node2vec and the BERT models to embed a knowledge graph with description information of every entity. Then, embedded nodes were indexed with a K-D tree structure. Finally, we used nearest neighbor search on K-D tree to retrieve neighbor-embedded nodes and implemented logical operations like projection, intersection, union, and negation to find answers to complex questions. We tested our model on three benchmark knowledge graph datasets and showed that QUERY2BERT significantly improved accuracy and speed compared to other state-of-the-art models.https://ieeexplore.ieee.org/document/10836698/BERTK-D treek-NNknowledge graph embeddingmulti-hop reasoninglogical query |
spellingShingle | Truong H. V. Phan Phuc do QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries IEEE Access BERT K-D tree k-NN knowledge graph embedding multi-hop reasoning logical query |
title | QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries |
title_full | QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries |
title_fullStr | QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries |
title_full_unstemmed | QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries |
title_short | QUERY2BERT: Combining Knowledge Graph and Language Model for Reasoning on Logical Queries |
title_sort | query2bert combining knowledge graph and language model for reasoning on logical queries |
topic | BERT K-D tree k-NN knowledge graph embedding multi-hop reasoning logical query |
url | https://ieeexplore.ieee.org/document/10836698/ |
work_keys_str_mv | AT truonghvphan query2bertcombiningknowledgegraphandlanguagemodelforreasoningonlogicalqueries AT phucdo query2bertcombiningknowledgegraphandlanguagemodelforreasoningonlogicalqueries |