KG-Hopper: Empowering Compact Open LLMs with Knowledge Graph Reasoning via Reinforcement Learning
View PDF HTML (experimental) Abstract:Large Language Models (LLMs) demonstrate impressive natural language capabilities but often struggle with knowledge-intensive reasoning tasks. Knowledge Base Question Answering (KBQA), which leverages structured Knowledge Graphs (KGs) exemplifies this challenge due to the need for accurate multi-hop reasoning. Existing approaches typically perform sequential reasoning steps guided by predefined pipelines, restricting flexibility and causing error cascades due to isolated reasoning at each step. To address these limitations, we propose KG-Hopper, a novel Reinforcement Learning (RL) framework that empowers compact open LLMs with the ability to perform integrated multi-hop KG reasoning within a single inference round. Rather than reasoning step-by-step, we train a Reasoning LLM that embeds the entire KG traversal and decision process into a unified ``thinking'' stage, enabling global reasoning over cross-step dependencies and dynamic path exploration with backtracking. Experimental results on eight KG reasoning benchmarks show that KG-Hopper, based on a 7B-parameter LLM, consistently outperforms larger multi-step systems (up to 70B) and achieves competitive performance with proprietary models such as GPT-3.5-Turbo and GPT-4o-mini, while remaining compact, open, and data-efficient. The code is publicly available at: this https URL. Comments: Accepted to IJCNN 2026 Subjects: Computation and Language (cs.CL); Artificial Intelligence (cs.AI) Cite as: arXiv:2603.21440 [cs.CL] (or arXiv:2603.21440v4 [cs.CL] for this version) https://doi.org/10.48550/arXiv.2603.21440 arXiv-issued DOI via DataCite Submission history From: Shuai Wang [view email] [v1] Sun, 22 Mar 2026 23:07:29 UTC (295 KB) [v2] Fri, 27 Mar 2026 09:53:48 UTC (295 KB) [v3] Mon, 30 Mar 2026 14:46:23 UTC (295 KB) [v4] Tue, 14 Apr 2026 17:49:12 UTC (311 KB)
No replies yet. Be first.