Knowledge Retrieval and RAG Reinforcement Learning for LLMs

Does supervising retrieval steps outperform final answer rewards?

Can intermediate feedback on retrieval decisions—which documents to fetch, when to stop—train agentic RAG systems more effectively than rewarding only the final answer? This matters because poor retrieval paths can accidentally succeed or good ones can fail on noisy metrics.

Note · 2026-02-22 · sourced from RAG
RAG How should researchers navigate LLM reasoning research?

Agentic RAG systems must make sequences of retrieval decisions — which query to issue next, which documents to process, when to stop retrieving. Training these systems on final answer accuracy alone (outcome-only reward) evaluates the end result without supervising the path. Poor intermediate retrieval decisions can accidentally produce correct final answers; good decisions can be penalized by noisy evaluation metrics.

RAG-Gym demonstrates that fine-grained process supervision — providing reward signals for individual intermediate retrieval steps, not just the final answer — substantially boosts agentic RAG performance. The improvement comes from two directions: correct retrieval steps are explicitly rewarded, and incorrect steps (retrieving irrelevant documents, issuing redundant queries) are explicitly penalized.

Three post-training algorithms were compared: PPO, DPO, and online DPO. DPO with both positive and negative feedback significantly outperforms PPO and single-direction training. The mechanism: DPO trains the model to prefer good retrieval chains over bad ones by directly contrasting them. Providing negative examples (what a bad intermediate step looks like) gives the model a gradient direction that outcome-only reward cannot supply.

The parallel to reasoning: Does failed-step fraction predict reasoning quality better? shows that in reasoning chains, intermediate step quality predicts final quality better than global features. RAG-Gym shows the same at the agentic level: retrieval step quality determines answer quality better than final-answer reward alone can capture.


Source: RAG

Related concepts in this collection

Concept map
17 direct connections · 159 in 2-hop network ·dense cluster

Click a node to walk · click center to open · click Open full network for a force-directed map

your link semantically near linked from elsewhere
Original note title

process-level supervision substantially outperforms outcome-only reward for training agentic rag systems