Can externalizing reasoning into knowledge graphs help smaller models compete?
Can structuring LLM reasoning as explicit knowledge graph triples enable smaller, cheaper models to solve complex tasks more effectively? This matters because it could make advanced reasoning accessible without scaling model size.
Knowledge Graph of Thoughts (KGoT) proposes that instead of keeping reasoning internal to the model, LLM "thoughts" should be converted into structured KG triples and stored in a graph database. The architecture iteratively constructs a knowledge graph from the task statement: at each step, the LLM generates intermediate insights ("thoughts"), converts them into triples (e.g., "Gollum (LotR)" → "interpreted by" → "Andy Serkis"), and stores them in a graph store that serves as an evolving structured knowledge base.
The results: KGoT achieves a 29% improvement in task success rates on the GAIA benchmark (Level 3 — highest difficulty) compared to Hugging Face Agents with GPT-4o mini. Small, cost-effective models can efficiently process the structured KG representation to achieve performance levels comparable to much larger counterparts.
The key architectural advantages:
Transparency: Unlike opaque monolithic LLM generations, every reasoning step is explicitly stored as triples. Biased inference steps can be identified by inspecting the graph. This addresses the explainability problem that Does chain of thought reasoning actually explain model decisions?.
Noise mitigation: New triples can be explicitly checked for information quality before integration, and existing triples can be removed if redundant. The graph provides a structured surface for quality control that internal reasoning traces lack.
Modularity: The architecture is extensible toward different graph query languages and tools (math solvers, web crawlers, Python scripts). Tool outputs are also converted to triples, creating a unified structured representation.
The fundamental move is "turning the unstructured into the structured" — converting unstructured data (websites, PDFs, model thoughts) into structured KG triples. This externalization of reasoning into a persistent, queryable, inspectable structure is a distinct alternative to both internal CoT and multi-agent debate.
This connects to:
- Do chain of thought traces actually help humans understand reasoning? — KGoT resolves this decoupling by making the reasoning structure externally inspectable
- Can query-time graph construction replace pre-built knowledge graphs? — LogicRAG builds query-specific graphs; KGoT builds task-specific graphs; both are inference-time graph construction but for different purposes (retrieval vs. reasoning)
- Can reasoning topologies be formally classified as graph types? — KGoT is a concrete implementation of GoT-style reasoning with the addition of persistent storage and tool integration
Original note title
Externalizing reasoning into knowledge graph triples enables small models to solve complex tasks at a fraction of large model cost