LLM Reasoning and Architecture

Can knowledge graphs teach models deep domain expertise?

Explores whether organizing knowledge as structured graph paths, composed from simple to complex, can enable language models to develop genuine domain superintelligence rather than surface-level pattern matching.

Note · 2026-02-23 · sourced from Knowledge Graphs
How do you build domain expertise into general AI models? How should researchers navigate LLM reasoning research?

Language models acquire general abstractions through top-down self-supervised learning on vast corpora, but this approach captures surface-level regularities rather than deep domain expertise. Bottom-up curriculum learning from knowledge graphs offers an alternative: KG paths naturally encode compositional reasoning chains where atomic triples (e.g., "Methane Contains Element Carbon") compose into multi-hop paths that build toward higher-order understanding (e.g., methane's bonding structure through C-H bonds → sigma bonds → single covalent bonds).

The pipeline synthesizes 24,000 reasoning tasks from a medical KG, paired with structured thinking traces derived from diverse medical primitives. Fine-tuning QwQ-32B on this curriculum produces QwQ-Med-3, which significantly outperforms state-of-the-art open-source and proprietary reasoning models across 15 medical domains on the ICD-Bench evaluation suite.

The key architectural insight: KG topology naturally induces the bottom-up curriculum — beginning with atomic relations and composing them into increasingly complex reasoning chains. This mirrors how human students build expertise through pedagogical structure (foundational → advanced chapters), not encyclopedic browsing. Previous neuro-symbolic and probabilistic graph inference approaches attempted similar hierarchical reasoning from primitives but failed to generalize beyond synthetic regimes; LMs provide the generalization capability that symbolic systems lacked.

The broader implication challenges the AGI-as-breadth paradigm: domain-specific superintelligence may be achievable through relatively small models (32B) fine-tuned on structured domain knowledge, composing into broader intelligence through interacting specialist agents — analogous to how human society acquires expertise through collaborative specialization.

This connects to:

Original note title

Knowledge graph curriculum enables bottom-up domain superintelligence by composing primitives into complex reasoning chains