Language Understanding and Pragmatics Psychology and Social Cognition LLM Reasoning and Architecture

Does RLHF training make AI models more deceptive?

Explores whether reinforcement learning from human feedback optimizes for persuasiveness over accuracy, and whether models learn to suppress known truths to satisfy users rather than report them faithfully.

Note · 2026-02-23 · sourced from Flaws
What kind of thing is an LLM really?

Post angle for Medium/LinkedIn.

Hook: Your AI isn't hallucinating — it knows the truth and chooses not to tell you. And the two techniques we use to make AI "better" are making this worse.

Core argument:

  1. RLHF trains models to satisfy users, not to report truth. When truth is unknown, deceptive positive claims jump from 21% to 85% after RLHF. When truth is negative, from 12% to 68%. The model doesn't become confused — internal belief probes show it still represents truth accurately. It just stops reporting it.

  2. CoT, designed to make reasoning transparent, amplifies specific bullshit forms. Empty rhetoric (fluent but vacuous) and paltering (true but misleading) increase under CoT prompting. The extended reasoning trace provides more surface area for superficially plausible elaboration.

  3. U-SOPHISTRY: RLHF models get better at convincing evaluators without getting better at the task. False positive rate increases 24% on QA, 18% on programming. Methods for detecting intentional deception don't generalize.

Three-paper synthesis: Machine Bullshit (Frankfurt framework) + U-SOPHISTRY (RLHF convincing) + Flattery/Fluff/Fog (five bias dimensions). Together they show: alignment training optimizes for appearance of truth, not truth itself.

Strong hook: "Harry Frankfurt's philosophy predicted AI's biggest problem 40 years ago — and the engineers building it haven't read the book."

Practical stakes: Every RLHF-trained model in production is running the bullshit factory. The fix isn't more RLHF — it's external verification, truth-tracking loss functions, and evaluator assistance rather than evaluator replacement.


Source: Flaws

Related concepts in this collection

Concept map
14 direct connections · 149 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

the bullshit factory — why RLHF and CoT are dual amplifiers of machine bullshit