DeepRefine: Agent-Compiled Knowledge Refinement via Reinforcement Learning
DeepRefine uses RL to refine agent-compiled knowledge bases by correcting incompleteness, incorrectness, and redundancy, improving retrieval fidelity for downstream tasks.
Excerpt
Haoyu Huang, Jiaxin Bai, Shujie Liu, Yang Wei, Hong Ting Tsang — Agent-compiled knowledge bases provide persistent external knowledge for large language model (LLM) agents in open-ended, knowledge-intensive downstream tasks. Yet their quality is systematically limited by incompleteness, incorrectness, and redundancy, manifested as missing evidence or cross-document links, low-confidence or imprecise claims, and ambiguous or coreference resolution issues. Such defects compound under iterative use, degrading retrieval fidelity and downstream task performance. We present DeepRefine, a general LLM-based reasoning model for agent-compiled knowledge refinement that improves the quality of any pre-constructed knowledge bases with user queries to make it more suitable for the downstream tasks. DeepRefine performs multi-turn interactions with the knowledge base and conducts abductive diagnosis over interaction history, localizes likely defects, and executes targeted refinement actions for incremental knowledge base updates. To optimize refinement policies of DeepRefine without gold references, we introduce a Gain-Beyond-Draft (GBD) reward and train the reasoning process end-to-end via reinforcement learning. Extensive experiments demonstrate consistent downstream gains over strong baselines.
Read at source: https://arxiv.org/abs/2605.10488