What makes delegation work beyond just splitting tasks?
Delegation is more than task decomposition. What dimensions of a task—like verifiability, reversibility, and subjectivity—determine whether an agent can safely and effectively handle it?
Current multi-agent frameworks treat delegation as task decomposition: split a goal into subtasks, assign them to agents, collect results. This is necessary but insufficient. Delegation in the full sense involves the transfer of authority, assignment of responsibility, implication of accountability for outcomes, risk assessment moderated by trust, capability matching, continuous performance monitoring, dynamic adjustments based on feedback, and ensuring completion under specified constraints.
The framework identifies 11 task characteristic axes that determine how delegation should be designed:
- Complexity — difficulty level, correlated with sub-steps and reasoning sophistication
- Criticality — importance and severity of failure consequences
- Uncertainty — ambiguity regarding environment, inputs, or success probability
- Duration — time-frame from instantaneous sub-routines to weeks-long processes
- Cost — computational expense including token usage, API fees, energy
- Resource requirements — specific tools, data access, human capabilities needed
- Constraints — operational, ethical, or legal boundaries limiting the solution space
- Verifiability — difficulty and cost of validating outcomes; high verifiability (formal proofs, code verification) enables "trustless" delegation; low verifiability (open-ended research) requires high-trust delegatees or expensive oversight
- Reversibility — whether effects can be undone; irreversible tasks (financial trades, database deletion) require stricter liability firebreaks than reversible tasks (drafting an email)
- Contextuality — volume and sensitivity of required external state; high-context tasks introduce larger privacy surfaces
- Subjectivity — whether success criteria are preference-based or objective; highly subjective tasks require human value specification and iterative feedback loops
Four design variables complete the framework: granularity (fine-grained vs coarse-grained objectives, where coarse requires further decomposition by delegatee), autonomy (full autonomy vs prescriptive specification), monitoring (continuous, periodic, or event-triggered), and reciprocity (whether delegation is one-way or mutual in collaborative agent networks). Reciprocity matters because in multi-agent systems, agents may delegate to each other — the delegation relation is not always hierarchical.
Verifiability determines whether evaluation is possible at all. The verifiability axis is not only a delegation criterion — it is a precondition for evaluation itself. Subjective, non-verifiable work (aesthetic judgment, strategic framing, interpretive analysis) leaves AI performance assessable only through proxies: style, fluency, apparent comprehensiveness, polish. Since Does polished AI output trick audiences into trusting it?, these proxies are exactly the dimensions AI is optimized to satisfy. The result is structural: as AI handles more subjective work, evaluation becomes harder rather than easier, because the only signals available to the evaluator are signals the system was trained to maximize. Delegation on low-verifiability axes is therefore not just a trust problem — it is a measurement problem where the instruments available all read positive.
The most design-relevant axes are verifiability, reversibility, and subjectivity because they determine the delegation contract — not just what to delegate, but how much trust, oversight, and rollback capacity the system needs. Since Does structured artifact sharing outperform conversational coordination?, the delegation framework provides the principled basis for what those SOPs should specify.
Source: Agents Multi Architecture
Related concepts in this collection
-
Does structured artifact sharing outperform conversational coordination?
Explores whether agents coordinating through standardized documents rather than natural language messages achieve better collaboration outcomes. Matters because it challenges the default conversational paradigm in multi-agent system design.
MetaGPT: SOPs as coordination mechanism; the delegation framework specifies what the SOP content should address
-
Can extreme task decomposition enable reliable execution at million-step scale?
Can breaking tasks into maximally atomic subtasks with voting-based error correction solve the fundamental reliability problem in long-horizon tasks? This challenges whether better models or better decomposition is the path to high-reliability AI systems.
MAKER: extreme decomposition works for fine-grained parallelizable tasks; the delegation framework identifies when fine vs coarse granularity is appropriate
-
How do agentic AI systems decompose into adaptation paradigms?
What are the core dimensions that distinguish different approaches to adapting agents and tools in agentic systems? Understanding this taxonomy could clarify which adaptation strategy fits which problem.
adaptation taxonomy; delegation adds trust, accountability, and reversibility dimensions
-
Why do AI agents fail at workplace social interaction?
Explores why current AI agents struggle most with communicating and coordinating with colleagues in realistic workplace settings, despite strong reasoning capabilities in other domains.
real-world delegation failure; the 11 axes explain which task properties create the hardest delegation challenges
-
Can AI guidance reduce anchoring bias better than AI decisions?
When humans and AI collaborate on decisions, does providing interpretive guidance instead of proposed answers reduce both over-trust in machines and abandonment on hard cases?
LTG adds "guide" as a third delegation mode beyond "automate" and "defer": for tasks high on subjectivity, irreversibility, and accountability — the axes where full delegation is most dangerous — the machine highlights useful aspects rather than proposing decisions
-
When should human-agent systems ask for human help?
Explores the timing problem in collaborative AI systems: since there's no objective metric for optimal interruption, how can we design deferral mechanisms that know when to involve humans without constant disruption or silent failures?
Magentic-UI's six mechanisms (co-planning, co-tasking, action guards, verification, memory, multitasking) operationalize the delegation framework: action guards address irreversibility, co-tasking addresses the verifiability challenge, and the absence of ground truth for deferral timing reflects the subjectivity axis
Click a node to walk · click center to open · click Open full network for a force-directed map
Original note title
intelligent delegation requires eleven task characteristic axes beyond decomposition — verifiability reversibility and subjectivity determine delegation design