NOV12
WED2025

Intelligence Increase as Control Under Uncertainty

Why IntelligenceargmaxπE[tγt1R(st,at)]\operatorname{Intelligence} \approx \arg\max_\pi \mathbb{E}[\sum_t \gamma^{t-1} R(s_t, a_t)] under tool-augmented constraints.
aiagentscontrolphilosophysmilepompdmcpplotlysystems

Timothy Leary’s “Intelligence Increase” has the vibe of a mystical unlock. You can restate it in a much more boring, much more useful way:

Intelligence is the ability of an agent to select actions that drive the world into high-value states, under uncertainty and resource constraints.

Once you phrase it like that, the whole modern AI stack — language models, MCP-style tool servers, agents interacting with infrastructure and robots — looks like a very particular answer to a very old control problem.

This post is about that control problem:

  • how to model it,
  • how modern agents and tools fit into it,
  • why “touching the physical world” is just the same structure with messier physics,
  • and where “intelligence increase” shows up in actual numbers.

Along the way, we’ll use a few simple visuals and mental models to keep the math grounded.


The basic object: an agent in a messy world

We’ll start with a standard partially observable Markov decision process (POMDP):

  • Hidden state: stSs_t \in \mathcal{S}
  • Observation: otO(st)o_t \sim O(\cdot \mid s_t)
  • Action: atAa_t \in \mathcal{A}
  • Transition: st+1T(st,at)s_{t+1} \sim T(\cdot \mid s_t, a_t)
  • Reward: rt=R(st,at)r_t = R(s_t, a_t)

An agent is a policy π\pi that picks actions from the full history:

atπ(ht),ht=(o1,a1,,ot1,at1,ot).a_t \sim \pi(\cdot \mid h_t), \quad h_t=(o_1,a_1,\dots,o_{t-1},a_{t-1},o_t).

The agent’s “intelligence,” in this formal sense, is how high it can drive the expected return:

Eπ[t=1Tγt1rt],\mathbb{E}_\pi \left[ \sum_{t=1}^{T} \gamma^{t-1} r_t \right],

subject to constraints: computation, information, safety, and so on. Here γ(0,1]\gamma \in (0,1] is the familiar discount factor from control theory: a knob that trades off patience vs. urgency by down-weighting far-future rewards (and keeps the infinite-horizon sum finite when γ<1\gamma<1). Thinking about “intelligence increase” is really thinking about how to make that discounted return larger under the same or tighter constraints.

When we talk about “intelligence increase” in 2025, we are really talking about pushing up this value for a whole distribution of tasks and agents:

  • better approximate world models T^,O^,R^\hat{T}, \hat{O}, \hat{R},
  • better policies π\pi that can use those models,
  • larger and more expressive action spaces A\mathcal{A} via tools and actuators.

That skeleton fits everything from pure software agents to robots and labs.

Policy π
Environment
emits
drives
influences
feedback
History h_t
Action a_t
Hidden state s_t
Observation o_t
Reward r_t

Language models as approximate world models

A large language model is, at minimum, a conditional distribution

pθ(xtx<t)p_\theta(x_t \mid x_{<t})

trained to minimize cross-entropy on sequences from some data distribution.

You can reinterpret this in control-theory language:

  • The history xtx_{\le t} is compressed into some latent state zt=fθ(xt)z_t = f_\theta(x_{\le t}).
  • The model learns an approximate predictive distribution over “what happens next,” conditioned on that latent.

If the text you train on includes:

  • bug reports and code reviews,
  • tickets and resolutions,
  • lab notebooks and experimental outcomes,
  • contracts and counteroffers,

then the model is a learned world model over human procedures. It’s not just predicting grammar; it is predicting how humans tend to evolve states in task-space.

Call that learned model P^θ\hat{P}_\theta. When you prompt it with:

“Given the following codebase and bug report, propose a fix and patch…”

you are implicitly querying

P^θ(‘good patch’‘bug + context’).\hat{P}_\theta(\text{‘good patch’} \mid \text{‘bug + context’}).

On its own, that is a powerful but passive object: a stochastic simulator of plausible next moves in human workflows.

To get agents, you wrap it in a control loop.

Predict
Sample or rank
History x<=t
Encoder f_theta
Latent state z_t
predict next
next continuation

From models to agents: policies, value, and tools

Imagine a simple agent built around an LLM:

  1. At time tt, history hth_t is encoded as a prompt xtx_t.
  2. The model samples or searches over continuations, producing a candidate action description a~t\tilde{a}_t.
  3. A parser maps a~t\tilde{a}_t into a concrete action atAa_t \in \mathcal{A}.
  4. The environment responds with a new observation and reward.

Formally, the policy is:

πθ(atht)=a~t1{parse(a~t)=at}  pθ(a~tprompt(ht))da~t.\pi_\theta(a_t \mid h_t) = \int_{\tilde{a}_t} \mathbf{1}\{\text{parse}(\tilde{a}_t)=a_t\} \\ \; p_\theta(\tilde{a}_t \mid \text{prompt}(h_t)) \, d\tilde{a}_t.

So far, the action space A\mathcal{A} is “emit text and hope someone else makes sense of it.” The intelligence increase here is limited by how many humans you can point at the outputs.

Now introduce tools.

Each tool kk is a conditional operator:

Tk:XkYk,T_k: \mathcal{X}_k \rightarrow \mathcal{Y}_k,

which might be:

  • a database query,
  • a code-execution sandbox,
  • an internal HTTP API,
  • a job scheduler,
  • or a robot controller.

The agent can choose between:

  • a language action (produce text), and
  • a tool action specifying (k,xk)(k, x_k) — “call tool kk with input xkx_k”.

The action space becomes:

A=Atext    k{(k,xk):xkXk}.\mathcal{A} = \mathcal{A}_{\text{text}} \;\cup\; \bigcup_k \{ (k, x_k) : x_k \in \mathcal{X}_k \}.

The Model Context Protocol (MCP) is a way to turn these tools into typed, discoverable, permissioned endpoints. From the agent’s perspective, MCP turns “the world of tools” into a graph of callable stochastic operators with schemas.

Mathematically, nothing mystical happens. You have:

  • expanded the set of admissible actions,
  • introduced new observations (tool outputs),
  • given the agent more ways to influence the environment.
Agent policy π
Text action
Tool action
Plan / prompt
Choose action
Emit text
Pick tool
mcp://repo
mcp://crm
mcp://infra
mcp://warehouse
mcp://lab
Result / observation

Digital vs physical: same math, uglier kernel

From the agent’s point of view, there is no metaphysical difference between:

  • calling a repo tool to refactor a codebase, and
  • calling a robot to move a box in a warehouse.

Both are just actions with stochastic consequences:

st+1T(st+1st,at).s_{t+1} \sim T(s_{t+1} \mid s_t, a_t).

The difference is the shape of TT.

For purely digital tools (say deterministic code transforms or database reads), the transition kernel is almost deterministic and cheap:

st+1f(st,at)with low variance and short latency.s_{t+1} \approx f(s_t, a_t) \quad \text{with low variance and short latency.}

For physical tools (robots, lab equipment, vehicles), you get:

st+1Tphysics(st,at)s_{t+1} \sim T_{\text{physics}}(\cdot \mid s_t, a_t)

with:

  • heavy-tailed noise (slip, friction, collisions),
  • partial observability (occlusions, sensor limits),
  • latency and dead time,
  • nonstationarities (wear, temperature, changing layout).

From the policy’s eye-view:

  • The expanded action set looks the same — “I can call move_box() just like I can call run_migration().”
  • The risk and uncertainty profiles are different — one lives in a low-noise, reversible subspace; the other in a high-noise, sometimes-irreversible subspace.

Robotics, autonomous labs, self-driving fleets — all of these are the same control problem with nastier TT, not a fundamentally different intelligence problem.

Physical actions
Wide, heavy-tailed noise
Latency, dead time
Sometimes irreversible
Digital actions
Narrow, low-variance transitions
Fast, reversible
Same policy class π

Why agents + tools feel like a phase transition

The last few years delivered three coupled changes:

Better approximate Bayesian predictors

LLMs are not exact Bayes filters, but empirically they are strong estimators of

p(xt+1xt)p(x_{t+1} \mid x_{\leq t})

over human-generated trajectories.

That gives agents cheap access to powerful priors over what tends to work next in human task-space. Instead of planning from scratch, they can sample from “what competent humans do” and refine from there.

Larger, more structured action spaces (via MCP)

Each new MCP server adds a set of typed tools:

  • mcp://repo → code and version control,
  • mcp://infra → deployment, scaling, configuration,
  • mcp://crm → customer records, workflows,
  • mcp://warehouse → picking and routing,
  • mcp://lab → experiment design and execution.

In the abstract, each of these is a bundle of new actions and transition dynamics stitched into the global kernel TT.

Awith MCPAtext only\mathcal{A}_{\text{with MCP}} \supset \mathcal{A}_{\text{text only}}

The reachable state region Rπ\mathcal{R}_\pi — all states the agent can drive the system into within a horizon — gets bigger and richer.

Cheap, compositional planning

Because the agent has a world model that is “good enough,” it can do cheap approximate planning:

  • sample candidate action sequences (plans),
  • evaluate them with a value head or a secondary model,
  • pick the best one under a heuristic return estimate.

This is Monte Carlo tree search with a learned policy/value prior, but the tree is spanned in natural language and tool calls instead of board moves.

Taken together:

  • the policy class πθ\pi_\theta we can realize got larger,
  • the learned dynamics model P^θ\hat{P}_\theta we can query got richer,
  • the action space A\mathcal{A} got deeper and more compositional.

“Intelligence increase” is not one knob; it’s a correlated push along all three axes.

Agent + MCP tools
Text-only agent
subset
Add tools + better policy
Larger region R_tools
Reachable region R_text

Callout: BCIs as just another channel

BCIs: extra I/O in the same POMDP

Brain–computer interfaces fit neatly into the same picture:

  • A recording BCI adds a new observation channel otBCIo_t^{\text{BCI}} encoding some function of neural state.
  • A stimulation BCI adds an action component atBCIa_t^{\text{BCI}} that influences the user’s brain state.

From the environment’s perspective, it is still:

st+1T(st+1st,at),at=(attools,atBCI).s_{t+1} \sim T(s_{t+1} \mid s_t, a_t), \quad a_t = (a_t^{\text{tools}}, a_t^{\text{BCI}}).

Today’s BCIs mostly restore missing channels (e.g. letting paralyzed people emit meaningful actions again). That is a genuine intelligence increase at the system level: the human-agent pair has a larger, more reliable action and observation space than the injured human alone.

Long term, you can treat BCIs as yet another class of MCP servers — wired into the nervous system instead of a database, warehouse, or robot arm.


Where “intelligence increase” shows up in numbers

All of this is nice concept art, but where do you actually see intelligence increase?

You can track it in at least four quantitative ways.

Value uplift on a task distribution

Fix some distribution over tasks (bugs, tickets, analyses, experiments). Let:

  • VhumanV_{\text{human}} = expected return with humans alone.
  • VaugV_{\text{aug}} = expected return with agents + tools + human oversight.

The first and most direct quantity is:

ΔV=VaugVhuman.\Delta V = V_{\text{aug}} - V_{\text{human}}.

For some domains, this looks like “higher resolution rate”; for others, “more revenue,” “fewer safety incidents,” or “more high-quality hypotheses per week.”

Human-only value V_h
ΔV = V_aug - V_h
Augmented value V_aug
Track uplift per task distribution

Cost per correct action (learning curve)

Let:

  • ChumanC_{\text{human}} be the cost per episode with humans alone.
  • CaugC_{\text{aug}} be the total cost (compute, infra, supervision) per episode with agents.

Then compare:

ChumanVhumanvsCaugVaug.\frac{C_{\text{human}}}{V_{\text{human}}} \quad \text{vs} \quad \frac{C_{\text{aug}}}{V_{\text{aug}}}.

As you accumulate more episodes, you often see a learning-curve-like decline in cost per correct action for the augmented system, roughly:

cost(cumulative successful actions)b.\text{cost} \propto (\text{cumulative successful actions})^{-b}.

When the augmented system wins that race, you have not only more capability but structurally cheaper capability.

Early episodes
High cost per correct action
More usage
Cost declines
Learning curve flattening
Cheaper per correct action
Compare vs human baseline
Augmented wins
Augmented loses

Share of agent-completed work

For a specific workflow, track:

  • the fraction of tasks completed end-to-end by agents (with human review), and
  • the rollback rate (cases where humans have to discard the agent’s work and redo it).

Plot the share of agent-completed work over time. You usually get an S-curve:

  • Assistant phase: agents help but rarely run workflows end-to-end.
  • Co-worker phase: agents own 20–50% of tasks, with humans mostly supervising.
  • Infrastructure phase: agent involvement is ubiquitous and boring, like databases.
Assistant phase
Agents assist
Co-worker phase
20-50% owned
Infrastructure phase
Agents ubiquitous

Predictive uncertainty

For a task-relevant function f(st+k)f(s_{t+k}) (e.g. “is the customer satisfied?”, “is the system in a safe state?”), look at:

Var[f(st+k)ht]\text{Var}\big[ f(s_{t+k}) \mid h_t \big]

under your best model. If better agents + tools steadily reduce that variance (and your calibration checks say it’s honest), then you have tighter control over futures that matter.

Better models, richer tools, and smarter policies all contribute to shrinking those error bars.


So what does “intelligence increase” amount to?

Strip away the acronyms and diagrams, and the picture is:

  • We’ve trained big sequence models that are surprisingly good at predicting what competent humans do next in a huge variety of contexts.
  • We’ve wired those models into MCP servers that expose databases, code, infrastructure, and increasingly, physical actuators.
  • We’ve wrapped the whole thing in agent loops that can plan, call tools, observe results, and adjust.

From the agent’s point of view, hitting a Kubernetes cluster and hitting a robot arm are just two different tools with different noise models. The math is the same; physics just makes the transition kernel uglier, slower, and more expensive to explore.

What people experience as “AI getting smarter” is:

  • policies π\pi that can exploit better world models,
  • over richer action spaces,
  • under more realistic constraints,
  • on more and more of the state space we care about.

Leary talked about Intelligence Increase like a moment of awakening. What we seem to be building is a layered control stack:

  1. Approximate world models over human behavior and environments,
  2. Agents that use those models to plan and choose actions,
  3. MCP tools that let those actions reshape both digital and physical reality.

The mathematics does not tell us what to value. It just gives us a sharper, cheaper way to steer whatever we decided to value in the first place. That is the real weight behind intelligence increase: not that we become godlike, but that we become dangerously competent at optimizing the particular reward functions we enshrine in code.

Humans
Goals, oversight
Agent policies π
MCP tools
APIs, infra, robots
Environment
Digital + physical state
Observations, rewards