Can computational power accelerate scientific discovery itself?
Does the pace of research breakthroughs scale with computing resources, like model performance does? ASI-ARCH tested this by running thousands of autonomous experiments to discover neural architectures.
ASI-ARCH is the first demonstration of fully autonomous neural architecture discovery at scale. A three-module agent system (Researcher → Engineer → Analyst) with persistent memory conducted 1,773 autonomous experiments over 20,000 GPU hours, discovering 106 state-of-the-art linear attention architectures.
The most significant finding is not the architectures themselves but the discovery of the first empirical scaling law for scientific discovery: architectural breakthroughs scale computationally. This transforms research progress from a human-limited process to a computation-scalable one.
Key mechanisms enabling this:
- Cognition base: ~100 seminal papers structured into scenario/algorithm/historical-context entries. The Analyst's problem summaries query this base by semantic similarity, ensuring relevant prior knowledge surfaces at the right time.
- Self-revision: Unlike previous NAS approaches that discard failed runs, ASI-ARCH requires agents to fix their own implementation errors. Iterative debugging continues until training succeeds — preserving promising ideas that would be lost to simple coding mistakes.
- Quality assurance: Real-time monitoring of training logs detects anomalies (excessive training duration, abnormally low loss) and terminates runs proactively, reporting issues for revision.
- Two-tier evolution: Parents from top-10 performers serve as bases for modification; references from positions 11-50 provide diversity. This balances exploitation with exploration.
- Novelty check: Embedding-based similarity search against historical motivations prevents redundant proposals.
The "AlphaGo Move 37" analogy is deliberate: AI-discovered architectures demonstrate emergent design principles that systematically surpass human baselines and illuminate previously unknown pathways — design insights invisible to human designers.
The exploration-then-verification two-stage strategy is practical: broad exploration on small models (cheap) narrows to rigorous validation on larger models (expensive). This is the same ladder-of-scales principle that makes research affordable.
ADAS as precursor (from Arxiv/Agents): The Automated Design of Agentic Systems (ADAS) research formulated the explicit thesis that "the history of machine learning teaches us that hand-designed solutions are eventually replaced by learned solutions." ADAS demonstrated that agents can be defined in code and new agents automatically discovered by a meta-agent programming ever better ones. While ASI-ARCH focuses on neural architecture search, ADAS extends the principle to agentic system design itself — the meta-agent discovers novel building blocks and combinations for agent architectures. This confirms the scaling law operates at multiple levels: not just architectures, but the design of the systems that discover architectures.
Source: Novel Architectures, Agents
Related concepts in this collection
-
What capabilities do AI systems need for autonomous science?
Explores whether current AI benchmarks actually measure what's required for independent scientific research—hypothesis generation, experimental design, data analysis, and self-correction—or if they test only adjacent skills.
ASI-ARCH implements all four: hypothesis generation, experimental design, data analysis, iterative self-correction
-
Do language models generate more novel research ideas than experts?
Explores whether LLMs can break free from expert constraints to generate more novel research concepts. Matters because novelty is often thought to be AI's creative blind spot.
ASI-ARCH's self-revision mechanism addresses the feasibility gap: ideas that fail are debugged, not discarded
-
Can human-AI research teams improve faster than autonomous AI systems?
Explores whether keeping humans actively involved in AI research collaboration accelerates paradigm discovery compared to fully autonomous self-improvement, and what safety advantages this preserves.
ASI-ARCH represents the fully autonomous end of this spectrum
-
Can AI systems discover better neural architectures than humans?
Can multi-agent LLM systems, when structured with genetic programming, discover novel neural network designs that outperform human-engineered architectures? This matters because it could automate a critical bottleneck in AI research.
Genesys: LLMs as evolutionary operators for NAS; complementary to ADAS meta-agent approach
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
autonomous architecture discovery follows an empirical scaling law — research breakthroughs are computationally scalable