Discovering Latent Facts from Context to Construct Richer Open Knowledge Graphs
DOI:
https://doi.org/10.1609/aaai.v40i23.38998Abstract
Knowledge graph construction (KGC) aims to extract valuable information from text and organize it into structured knowledge graphs (KGs). Recent methods have leveraged the strong generative capabilities of large language models (LLMs) to improve the generalization and reduce the labor costs. However, constrained by the input length of LLMs, existing methods mainly focus on extracting knowledge within individual texts and lack the capability to discover latent knowledge across texts. To fill this gap, we propose a novel method for open knowledge graph construction, termed KG-DLF. The core idea of this method is to enhance the knowledge graph construction process by discovering new facts that are consistent with the underlying contextual logic. Specifically, we first design a knowledge extractor to extract knowledge from the text. Then, a knowledge normalizer performs schema alignment on the extracted knowledge. Next, we explore a knowledge discoverer based on a clue search strategy, which leverages the logical consistency of context to mine latent facts. Finally, we design a counterfactual-based knowledge corrector, enabling the model to purify knowledge and reduce factual errors. Experimental results show that KG-DLF is capable of extracting comprehensive knowledge in open-world scenarios across three KGC benchmarks.Published
2026-03-14
How to Cite
Li, J., Yu, H., Ma, Z., & Qi, P. (2026). Discovering Latent Facts from Context to Construct Richer Open Knowledge Graphs. Proceedings of the AAAI Conference on Artificial Intelligence, 40(23), 19233–19241. https://doi.org/10.1609/aaai.v40i23.38998
Issue
Section
AAAI Technical Track on Knowledge Representation and Reasoning