Skip to content

Agents All the Way Down

Colony is built on a single, testable conjecture:

General intelligence is emergent from the right composition of LLM-based reasoning policy and action space.

General Intelligence Emerges from Composition

The action space is as important to foster intelligent behavior as the reasoning model itself. By providing the right set of actions, we can guide the LLM to both learn (at training time) and reason (at inference time) in more effective ways. This is corroborated by embodied AI research which shows that the action space available to an agent significantly influences its ability to learn and perform tasks.

The conjecture is bold: compose enough LLM-based agents with the right mix of capabilities, and general intelligence emerges. Colony is the testbed for that conjecture.

This architectural claim has specific consequences for how the framework is built.

Nested Action Policies

Add explanation here of how nested recursive action policies enable the emergence of general intelligence, but we can exploit agents to emulate this recursive structure.

Nested Action Policies

Add explanation here of how the action policy is an aspect weaver that decides control and data flow inside an agent.

The Virtual Agent

Here is Colony's most provocative architectural idea: a multi-agent system is not a collection of independent agents collaborating on a task. It is the different cognitive levels of a single virtual agent.

Consider how human cognition works at different levels:

Level Human Cognition Colony Implementation
L0: Reflexive Immediate reactions, pattern matching Rule-based guards, reactive policies
L1: Deliberative Goal-oriented planning, sequencing LLM-based action policies, plan generation
L2: Reflective Self-assessment, strategy revision Reflection capabilities, meta-reasoning agents
L3: Meta-cognitive Reasoning about reasoning itself Supervisor agents, capability orchestration

In Colony, each level can be implemented by different agents with different capabilities. The top-level agent has higher-level, more abstract capabilities (strategic planning, meta-reasoning). Lower-level agents have specialized, fine-grained capabilities (page analysis, code inspection, hypothesis testing). Together, they implement the cognitive architecture of a single virtual agent whose reasoning depth and breadth exceed what any individual agent could achieve.

graph TB
    subgraph "Virtual Agent"
        Meta[Meta-Cognitive Agent<br/>L3: Reasoning about reasoning]
        Meta --> Reflect1[Reflective Agent<br/>L2: Strategy revision]
        Meta --> Reflect2[Reflective Agent<br/>L2: Self-assessment]
        Reflect1 --> Delib1[Deliberative Agent<br/>L1: Code analysis plan]
        Reflect1 --> Delib2[Deliberative Agent<br/>L1: Test generation plan]
        Reflect2 --> Delib3[Deliberative Agent<br/>L1: Documentation plan]
        Delib1 --> Page1[Page Agent<br/>L0: Analyze module A]
        Delib1 --> Page2[Page Agent<br/>L0: Analyze module B]
        Delib2 --> Page3[Page Agent<br/>L0: Generate tests]
    end

What This Means in Practice

The "agents all the way down" philosophy produces concrete architectural decisions: 1. Dynamic hierarchies. The agent hierarchy is not fixed at design time. Agents spawn sub-agents and agent pools, form teams and coalitions, play games, and dissolve -- all decided at runtime by the action policy based on the task.