Language Understanding and Pragmatics LLM Reasoning and Architecture

Can large language models translate natural language to logic faithfully?

This explores whether LLMs can convert natural language statements into formal logical representations without losing meaning. It matters because faithful translation is essential for any AI system that reasons formally or verifies specifications.

Note · 2026-02-21 · sourced from Natural Language Inference
What kind of thing is an LLM really? Where exactly does language competence break down in LLMs? How should researchers navigate LLM reasoning research?

"Faithful Autoformalisation with LLMs" evaluates whether LLMs can translate natural language statements into formal logical representations (first-order logic, modal logic, higher-order logic). The central finding: LLMs can produce syntactically well-formed logical expressions but fail to produce semantically faithful ones. Form and content come apart.

The failure modes are not random — they cluster at precisely the points where natural language semantics are most structurally complex:

This is a structurally significant finding because autoformalisation is a direct test of the relationship between linguistic competence and logical competence. If LLMs can handle syntax-level structure (as they demonstrably can for many tasks) but fail when that structure requires semantic commitment, then their linguistic processing operates at a level that stops short of truth-conditional content.

The finding connects to the broader pattern that Can models pass tests while missing the actual grammar?. For autoformalisation, "correct output" would mean syntactically valid logic; "genuine linguistic generalisation" would require semantic faithfulness. LLMs achieve the former but not the latter.

This also extends Do language models actually use their encoded knowledge? — even when LLMs encode semantic properties, that encoding does not reliably translate into semantic commitment during generation. The representation has the information; the generation does not use it.

Structured semantics understanding vs. generation asymmetry: "Probing Structured Semantics Understanding and Generation" (2401.05777) confirms the generation < understanding asymmetry for formal languages. LLMs can interpret formal language (translate logical form to natural language question) more accurately than they can generate it (translate natural language to logical form). Furthermore, formal language complexity matters: lower formalization (closer to natural language, like KoPL) is easier for models; higher formalization (SPARQL, Lambda DCS) fails at entity grounding. This suggests the autoformalisation failure is graded — the farther the target representation from natural language surface form, the worse the semantic fidelity.

The practical implication: LLM-assisted formal verification, specification writing, or logical reasoning pipelines cannot trust LLM-generated logical forms without post-hoc semantic verification. The syntactic plausibility of the output masks the semantic errors.


Source: Natural Language Inference; enriched from LLM Architecture

Related concepts in this collection

Concept map
17 direct connections · 128 in 2-hop network ·medium cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

llms fail at faithful autoformalisation because they cannot translate natural language to logical representations without semantic loss