What blocks scaling from language models to autonomous agents?
If large language models excel at next-token prediction, why do they struggle with long-horizon goal-oriented tasks? This explores whether the bottleneck is model capacity or the environments used to train them.
Nex-N1's diagnosis is that the LLM-to-agent transition is blocked by a misalignment between LLM pretraining (myopic next-token prediction) and the long-horizon goal-oriented nature of agentic tasks — and that bridging this requires not better models but a new scale of interactive environments. Scarcity of diverse environments leaves models as "System 1" responders without "System 2" rigor; lack of realistic grounding produces hallucinated tool use and brittle error recovery.
The structural claim is that environments must scale on three orthogonal dimensions, and a deficit on any one ruins the resulting policy. Complexity comes from agent hierarchies — NexAU is a lightweight high-throughput runtime that decouples agent definition from execution, treating sub-agents and tools as interchangeable functional units in a recursive ReAct-like architecture. Diversity comes from automated synthesis — NexA4A generates diverse agent architectures and workflows from natural-language specifications rather than human-designed templates, breaking the dependency on hand-built environments. Fidelity comes from grounding — NexGAP integrates real Model Context Protocol (MCP) tools and information fusion, generating trajectories rooted in authentic latency, stochasticity, and feedback loops.
The orthogonality matters because earlier frameworks fail in characteristic ways: rigid graph-based orchestrators provide reliability but limit diversity; pure synthetic environments provide diversity but break on real execution. Treating environments as generative language specifications rather than static code is the move that lets all three axes scale together. The empirical signal — Nex-N1 outperforms SOTA open-source models and approaches frontier proprietary models on SWE-bench and τ2 — supports the thesis that the limiting reagent has been environments, not parameters.
This stands in productive tension with Can 78 demonstrations teach agency better than 10000?, which argues that strategic data curation beats environment-scale; the resolution is likely that environment richness sets a ceiling that curated data exploits, not a substitute for curation.
Source: Action Models
Related concepts in this collection
-
Can 78 demonstrations teach agency better than 10000?
Does agentic capability depend on data volume or curation quality? LIMI achieves 73.5% on AgencyBench with 78 samples versus 24-45% for models trained on 10K+, suggesting strategic demonstration design may matter far more than scale.
tension with: LIMI argues 78 curated demos beat data abundance; Nex-N1 argues environments are the limiting reagent and must scale. Both can be true if environment richness sets the curation ceiling.
-
Can agents learn beyond what their training data shows?
Explores whether supervised fine-tuning on expert demonstrations creates a hard ceiling on agent competence, or whether agents can generalize to scenarios their curators never captured.
complements: explains why diverse environments matter — curated demos cap exploration to what curators imagined; environment scaling breaks that ceiling.
-
Can you turn an LLM into an agent by just fine-tuning?
Explores whether upgrading language models to action-producing systems requires only model retraining or demands a broader pipeline transformation including data collection, grounding, integration, and safety evaluation.
extends: LAM defines the pipeline structure; Nex-N1 specifies what environment scaling must look like at the data-collection and grounding stages.
-
Can agent deployment itself generate training signals automatically?
Can we extract learning signals from the natural next-states that agents encounter during real deployment—user replies, tool outputs, test verdicts—rather than relying on separate annotation pipelines? This reframes how agents improve continuously.
complements: high-fidelity environments produce informative next-state signals; the value of next-state learning depends on environment fidelity.
-
Why does random tool sampling produce unrealistic synthetic training data?
Tool-calling datasets generated through random sampling and single-turn framing lack the complexity and coherence of real deployment. This explores what structural choices in data synthesis determine whether models can learn realistic tool composition.
exemplifies: ToolFlow is the diversity-and-fidelity argument applied to one specific data-generation pipeline.
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
agentic training requires environment scaling along three orthogonal dimensions — complexity diversity and real-world fidelity must scale together