Skip to content

Latest commit

 

History

History
188 lines (114 loc) · 27.3 KB

File metadata and controls

188 lines (114 loc) · 27.3 KB

TF-08: The Exploration-Exploitation Balance (Hypothesis + Discussion)

Actions don't merely select among outcomes — they generate information about how the environment responds to interventions. This fundamental capacity creates an irreducible tension: the agent must balance exploiting what it already knows against exploring to improve its model. The optimal policy integrates both through a unified objective that jointly maximizes value and causal information yield.

Epistemic status: The structural claim — that the optimal policy jointly maximizes value and causal information — is discussion-grade, well-supported by convergent results in Bayesian RL, active inference, and information-directed sampling. The specific form of $\lambda(M_t)$ is not derived from first principles within TFT but is supported by independent theoretical traditions.

Causal Information Yield (CIY)

TF-02 establishes that action-contingent observations carry interventional (Level 2) information that passive observation cannot provide. To quantify this, define the canonical causal information yield of an action $a$ given model state $M$:

[Definition (CIY-canonical)] $$\text{CIY}(a;, M) = \mathbb{E}{a' \sim q(\cdot \mid M)}!\left[D{\mathrm{KL}}!\left(P(o \mid do(a), M) ,|, P(o \mid do(a'), M)\right)\right]$$

where $q(\cdot \mid M)$ is a reference distribution over comparator actions (uniform, policy-induced, or task-specific). This measures how strongly the action changes the interventional distribution of outcomes relative to alternatives. At time $t$, the agent evaluates $\text{CIY}(a;, M_{t-1})$ for candidate actions $a \in \mathcal{A}$ to guide exploration.

Because it is an expectation of KL divergences, $\text{CIY} \geq 0$ by construction.

Dependence on the reference distribution $q$. The quantitative value of $\text{CIY}(a; M)$ depends on the choice of $q(\cdot \mid M)$, which is a significant degree of freedom. A uniform $q$ treats all alternative actions equally; a policy-induced $q$ emphasizes alternatives the agent would actually consider; a task-specific $q$ may weight high-stakes alternatives. The key question is whether CIY rankings — which action has highest CIY — are stable across reasonable choices of $q$. In general, they are not: an action that is maximally informative relative to a uniform baseline may not be maximally informative relative to the current policy's alternatives. TFT does not claim CIY rankings are $q$-invariant. Rather, the choice of $q$ is itself a design decision: it encodes "informative relative to what alternatives?" When CIY appears in downstream objectives, the associated $q$ should be specified.

Default convention for empirical work. TFT adopts the policy-induced $q$ as default: $q(\cdot \mid M) = \pi(\cdot \mid M)$, the agent's current policy. This yields CIY as "how different is this action's outcome from what I'd typically see?" — the most decision-relevant comparison. When an alternative $q$ is used (uniform for worst-case analysis, task-specific for focused probing), it should be explicitly stated. CIY values are not comparable across different choices of $q$.

For diagnostic use with observational statistics, define a proxy:

[Definition — Auxiliary Proxy] $$\text{CIY}{\text{proxy}}(a{t-1}) = I(o_t; a_{t-1} \mid M_{t-1}) - I(o_t; a_{t-1} \mid \Omega_t, M_{t-1})$$

This proxy can be useful but is sign-indefinite in general and requires causal assumptions for interpretation. TFT treats $\text{CIY}$ (interventional definition) as canonical; $\text{CIY}_{\text{proxy}}$ is auxiliary.

Sign-indefiniteness and the observation model. The second MI term $I(o_t; a_{t-1} \mid \Omega_t, M_{t-1})$ measures dependence between action and observation through the observation mechanism itself (TF-01's action-dependent $h$). Under the action-independent special case $o_t = h(\Omega_t, \varepsilon_t)$, this term vanishes — conditioning on $\Omega_t$ screens off $a_{t-1}$ — and the proxy reduces to $I(o_t; a_{t-1} \mid M_{t-1}) \geq 0$. Sign-indefiniteness arises specifically when actions affect the observation mechanism (sensor selection, query choice, attentional direction): the second term can then exceed the first through an explaining-away structure (conditioning on the full state can increase the apparent action-observation dependence relative to the model-conditional baseline). When using the proxy diagnostically, note which regime applies.

Why this is non-trivial: $\text{CIY} = 0$ for a passive observer (no action variation, or action with no effect on outcome distribution). $\text{CIY} > 0$ when actions causally alter what is observed. This is exactly what distinguishes Level 2 (interventional) from Level 1 (associational) epistemic access.

Maximizing causal information yield is exactly what good exploration does — the "investigating" function of Feldbaum's dual control1: choosing actions whose consequences are maximally informative about the causal structure of the environment. Note that the "environment" may include other agents whose models can be queried. When a query action elicits a response from a knowledgeable source, the CIY can be extremely high — the response is causally downstream of the query and carries pre-compressed information about the environment that would require many direct probes to reconstruct. See the Query Actions section below.

CIY Admissibility Regimes

To use CIY rigorously in the unified policy objective (below) and in exploration arguments throughout, we distinguish three regimes:

Regime A — Randomized interventions. When actions are randomized (or otherwise exogenous conditional on $M_{t-1}$), the interventional distributions $P(o_t \mid do(a), M_{t-1})$ are directly estimable from observed outcomes. This is the standard case for active agents performing exploration: RL agents trying different actions, scientists running randomized experiments, organisms probing their environment. CIY is directly estimable and non-negative in this regime. The proxy $\text{CIY}_{\text{proxy}}$ is also easiest to interpret here.

Regime B — Observational with causal assumptions. When the agent cannot freely vary its actions (policy-constrained, nearly passive, or observational), CIY estimation requires additional causal assumptions: a known DAG structure, instrumental variables, or functional form assumptions about confounding. In this regime, CIY estimates carry an identifiability qualifier — they are valid only to the extent that the causal assumptions hold. CIY is conditionally estimable; results that depend on it inherit the causal assumptions.

Regime C — Adversarial communication. When the observation channel includes responses from potentially adversarial sources, the agent's own CIY (from its action of querying) remains non-negative — the query causally generates a response. However, the content of that response may be designed to increase the agent's model-reality mismatch rather than decrease it. This is not "negative CIY" — the mutual information between query and response is still positive — but rather adversarial exploitation of the agent's update channel. The adversary injects disturbance through $\rho$ by leveraging the agent's trust (high $\eta^*$) to drive a large, misdirected model update. See the Adversarial Mirror section below and Appendix A's coupling coefficient $\gamma_A$ for the formal treatment. CIY remains non-negative; the adversarial effect operates through the disturbance term, not through the information measure.

For TFT, downstream equations treat $\text{CIY}$ as the canonical interventional quantity (non-negative by construction). $\text{CIY}_{\text{proxy}}$ may be used as an estimator or heuristic in Regimes A/B, with explicit identifiability qualifiers.

Identifiability Gate: When to Use CIY

Before incorporating CIY into policy objectives or empirical analyses, verify:

  1. Action variation exists. The agent must actually vary its actions across comparable model states. An agent that always takes the same action in a given state has CIY $\approx 0$ for decision purposes — there is no interventional contrast.
  2. Regime is identified. Regime A (randomized/varied actions) permits direct CIY estimation. Regime B (observational) requires explicit causal assumptions — state these. Regime C (adversarial sources) requires separating the information channel from the disturbance channel.
  3. Reference distribution $q$ is specified. CIY values are not comparable across different $q$ choices. Use the policy-induced default unless otherwise justified (see above).
  4. Stationarity holds locally. CIY estimation requires that the model state $M_{t-1}$ and the environment dynamics are approximately stationary over the estimation window. Under rapid drift, CIY estimates are stale.

If any of these conditions fail, CIY-based exploration terms in the policy objective should be dropped or replaced with simpler uncertainty-based heuristics (e.g., UCB-style bonuses based on visit counts or ensemble disagreement).

CIY Identifiability

The canonical CIY definition uses interventional outcome distributions. This raises the practical question: when can those distributions be estimated from observable quantities? (For the proxy, additional dependence on latent $\Omega_t$ introduces further identifiability requirements.)

With interventional data (the typical case for active agents). An agent that varies actions across comparable model states and observes resulting outcomes can estimate $P(o_t \mid do(a), M_{t-1})$ empirically, and therefore estimate CIY directly via the KL expression above. This is the standard case in RL exploration, controlled experimentation, and active sensing.

With observational data only (passive or constrained agents). When the agent cannot freely vary actions, interventional distributions are not directly observed. CIY estimation then requires additional assumptions. Specifically, it requires either:

  • A causal graph (DAG) with known structure, enabling do-calculus adjustment
  • Instrumental variables that affect actions but not outcomes except through the causal pathway
  • Assumptions about the functional form of confounding

Without such assumptions, CIY is not identifiable from observational data alone. This is Pearl's fundamental insight restated in our formalism: you cannot learn causal structure from correlations without either experiments or causal assumptions. In this setting, $\text{CIY}_{\text{proxy}}$ is best treated as a diagnostic statistic, not a causally identified quantity.

Practical implication. Active agents have access to interventional data as a consequence of acting. The quality of CIY estimation depends on action diversity (exploration) and local stationarity of model state during estimation. An agent that always takes the same action in a given state cannot identify action-conditioned effects and effectively has CIY near zero for decision purposes. This provides an information-theoretic argument for exploration that complements TF-05's mismatch argument.

The Unified Policy Objective

The exploration-exploitation tension suggests a single objective that the optimal policy $\pi^*$ maximizes:

[Discussion — Normative Objective] $$\pi^*(M_t) = \arg\max_a \left[\mathbb{E}[\text{value}(a) \mid M_t] + \lambda(M_t) \cdot \text{CIY}_q(a;, M_t)\right]$$

The first term is the exploitation objective — expected value given the current model. The second term is the exploration objective — the causal information yield of the action (defined above as an expectation over the comparator distribution $q$; no additional expectation is needed since $\text{CIY}_q(a; M_t)$ is deterministic given $a$ and $M_t$). The coefficient $\lambda(M_t)$ controls the balance and should depend on the agent's epistemic state:

  • When $U_M$ is high (model uncertain): $\lambda$ is large — exploration is valuable because the model has much to learn.
  • When $U_M$ is low (model confident): $\lambda$ is small — exploitation dominates because the model is (probably) already good.
  • When the time horizon is long: $\lambda$ should be larger — the information gained now compounds over many future decisions.
  • When $\rho$ is high (fast-changing environment): $\lambda$ should be larger — the model is perpetually uncertain because the environment keeps changing (connecting to TF-10).

Note on dimensional consistency and status. The two terms in this objective have different natural units: the first is in value units (reward, utility, cost), the second in information units (bits, nats). The coefficient $\lambda(M_t)$ must therefore absorb the conversion — it carries units of [value per unit information] and effectively prices exploration in value-equivalent terms. This objective is normative (a design criterion for what the optimal policy should maximize) rather than descriptive (a claim about what agents actually compute). TFT identifies the structural form of the trade-off; $\lambda$'s magnitude is domain-specific and not derivable from TFT alone. However, $\lambda$ is not entirely unconstrained — in specific domains it reduces to known, sometimes derived quantities:

Domain What $\lambda$ reduces to Status
Bayesian bandits Gittins index (implicit information price derived from dynamic programming) Exactly derived
Kalman dual control Probing cost in expected quadratic objective (both terms in cost units) Exactly derived
Active inference Precision on epistemic affordance (both terms in free-energy units) Framework-derived
Information-directed sampling Ratio of squared value-of-information to information gain Exactly derived (Russo & Van Roy)
RL with UCB Confidence-bound scaling $c \sqrt{\ln t / N(a)}$ Heuristic (tuned)
Human decision-making Not explicit; manifests as curiosity drive vs. reward seeking Empirical

The pattern: $\lambda$ is derivable when the problem has sufficient structure (finite state/action, known dynamics, well-defined horizon). In less structured problems, $\lambda$ becomes a design parameter or emergent property. TFT's contribution is not to derive $\lambda$ universally but to identify the structural role it plays across all these settings. The CIY term assumes the agent operates in Regime A or B (see CIY Admissibility Regimes above).

Note on CIY estimability. This objective requires the agent to evaluate $\text{CIY}_q(a; M_t)$, which in turn requires either interventional data (Regime A) or causal assumptions (Regime B). An agent that cannot estimate CIY — because it lacks action variation, operates in a purely observational mode, or faces unidentifiable confounding — can still use the exploitation term alone; the exploration term becomes relevant only when CIY estimation is feasible. In practice, for active agents (RL, experimental science, active sensing), CIY is estimable by construction because the agent generates interventional data through its actions. For weaker-coupling agents (passive observers, constrained policy), the objective degrades gracefully to exploitation-only.

Connection to active inference. The Free Energy Principle's "expected free energy" objective decomposes into an extrinsic value term (pragmatic, goal-directed) and an epistemic value term (information-seeking). The TFT formulation above is structurally isomorphic: expected value ≈ extrinsic value, expected CIY ≈ epistemic value. Whether this convergence is deep (both are instances of the same mathematical principle) or superficial (similar-looking objectives from different foundations) is an open question. The TFT formulation has the advantage of grounding exploration in explicitly causal information rather than in entropy reduction, which may be more precise — not all uncertainty reduction is equally valuable; causal information is specifically what enables better intervention, not merely better prediction.

The Exploration-Exploitation Trade-off

Action selection faces a fundamental tension:

Exploit: Choose $a_t$ to maximize immediate predicted value given $M_t$: [Definition — Objective Component] $$a_t^{\text{exploit}} = \arg\max_a \mathbb{E}[\text{value}(a) \mid M_t]$$

Explore: Choose $a_t$ to maximize expected information about model errors: [Definition — Objective Component] $$a_t^{\text{explore}} = \arg\max_a \mathbb{E}[I(\delta_{t+1}; \Omega_t \mid M_t, a)]$$

Exploitation uses the model as-is. Exploration tests the model.

The optimal balance depends on:

  • Model uncertainty ($U_M$ high → explore more)
  • Time horizon (long → invest in exploration early)
  • Cost of exploration (high → explore cautiously)
  • Mismatch history (persistent $\delta \neq 0$ → investigate the source)

This connects directly to the zero-mismatch ambiguity (TF-05): an agent that only exploits will tend toward confirmation bias — observing only what its model already explains. Exploration is the mechanism by which the agent actively tests its model, converting the ambiguous case (b) in TF-05 into genuine signal.

Action as Information Generation

Actions don't merely affect the environment — they generate information that passive observation cannot provide. TF-02 establishes three levels of epistemic access grounded in the causal structure of the feedback loop: associational (Level 1), interventional (Level 2), and counterfactual (Level 3).

Action selection is what makes Levels 2 and 3 available. By choosing to act and then observing consequences, the agent generates causal information yield (defined above) — information about how the environment responds to interventions, not merely about correlations. This is why the feedback loop (action → observation → update) is more powerful than passive observation (observation → update) alone.

The exploration-exploitation trade-off (above) is fundamentally about how much causal information yield the agent seeks: exploitation maximizes predicted value at Level 1; exploration maximizes expected causal information yield at Level 2.

Query Actions: Accessing External Models

The discussion above — and the theory's examples throughout — has implicitly framed information-generating actions as direct environment probes: do something to the world, observe the result, update the model. But there is a qualitatively different class of actions with distinctive properties: querying another agent's model.

When a reliable external model exists — an expert, a database, a reference text, a trustworthy advisor, a well-trained LLM — the action "ask a well-formed question" can yield information equivalent to thousands of probe-observe cycles. The classic illustration: asked to measure a building's height with a barometer, one can drop it from the roof and time the fall, measure shadow ratios, swing it as a pendulum at the top and bottom — all Level 2 probes of the physical environment. Or one can offer the barometer to the building's janitor in exchange for the answer — accessing information that already exists compressed in another agent's model.

Why Query Actions Are Distinctive Within the CIY Framework

Information density. A single well-targeted query to a knowledgeable source can carry CIY orders of magnitude higher than any individual environment probe. The source's model has already performed the compression work (TF-03) — extracting the relevant sufficient statistic from a vast interaction history the querying agent has never had. The response transfers the output of that compression rather than requiring the agent to reconstruct it from scratch.

Trust-dependent gain. The update gain $\eta^*$ for query-derived information depends not on observation channel noise $U_o$ but on the agent's model of the source's model — a second-order epistemic state. How much should the agent update on the response? This depends on: the source's domain competence (does their model have high $S$ for the relevant quantity?), their reliability (are they truthful? calibrated?), and their alignment (do they share the agent's objectives, or might they be adversarial?). This connects to game theory and reputation as a meta-model — the agent needs a model of which sources to trust how much about what.

Pre-compressed information. Environmental probes return raw observations that must be compressed through the agent's own model. Query responses arrive already compressed in the source's representational framework. This is why they're so information-dense — but it also introduces a translation cost when the source's representation doesn't align with the agent's. An expert's answer may be incomprehensible to a novice not because the information is absent but because the agent lacks the representational capacity (TF-07's model class $\mathcal{M}$) to absorb it.

Structural adaptation via external models. Query actions can trigger not just parametric updates but structural change. Encountering another agent's model — through conversation, reading, apprenticeship, or consultation — is one of the primary mechanisms by which agents acquire new representational frameworks. This connects to TF-10's "grafting" mechanism: incorporating external structure rather than building it de novo. Boyd's thought experiment specifically illustrates cross-domain recombination — freeing components from their native domains and reassembling them into novel model structures.

Implications for optimal action selection. When high-CIY query channels are available, the unified policy objective (above) will tend to favor query actions over direct probes, particularly when:

  • The agent's own $U_M$ is high (it has much to learn)
  • A trusted source with high $S$ for the relevant domain is accessible
  • The cost of querying (social, monetary, time) is low relative to the cost of direct probing
  • The information needed is about structure (requiring many probes to reconstruct) rather than about the agent's specific situation (which only the agent can observe)

Direct probing remains essential when the information needed is situational (no external model has access to the agent's specific environment state), when no trustworthy source exists, or when the agent needs to verify claims rather than accept them. The optimal agent uses both channels, allocating actions to whichever has higher expected CIY per unit cost.

The Adversarial Mirror: Deception and Model Corruption

Query actions have a dark mirror in adversarial settings. The same communicative channel that enables cooperative information transfer — where one agent's response improves another's model — can be exploited to degrade the opponent's model. This is the domain of game theory, information warfare, and strategic communication.

Deception as adversarial disturbance injection. A cooperative query yields positive CIY — the response genuinely improves the agent's model. A deceptive response also yields positive CIY in the strict information-theoretic sense (the mutual information between query and response is still non-negative). But the content of the response is designed to increase rather than decrease model-reality mismatch: the agent updates its model in a direction that moves it away from the true environment state. The update gain $\eta^*$ for the victim depends on trust in the source; a successful deception exploits high trust to inject a large, misdirected update. In the Lyapunov framework (Appendix A), this is an adversary directly contributing to $\rho_B$ — not by changing B's physical environment, but by corrupting B's model of that environment. The adversary's communicative action functions as a disturbance $w(t)$ injected through the observation channel, with coupling coefficient $\gamma_A$ (Proposition A.3) determined by the victim's trust level and exposure to the adversary's signals.

Active OODA loop interference. A central theme in Boyd's work — and in strategic thought at least since Sun Tzu — is that an adversary can do more than merely adapt faster (TF-10); it can actively interfere with the opponent's feedback loop: generating ambiguous or contradictory signals to degrade the opponent's Orient phase, creating false patterns to induce model errors, manipulating the information environment to increase the opponent's $\rho$ while providing no genuine signal. Feints, disinformation, electronic warfare, strategic ambiguity, and propaganda are all instances of one agent using communicative or signal-manipulating actions to drive another agent's mismatch upward — the adversarial coupling $\gamma_A$ in Appendix A's Proposition A.3 operating through the information channel rather than through physical action.

Trust as a meta-model under adversarial pressure. In adversarial settings, the agent's model of source reliability becomes a critical adaptive target in its own right. An adversary who can corrupt this meta-model — making the victim trust unreliable sources or distrust reliable ones — achieves a second-order attack: not just injecting bad information, but compromising the victim's capacity to evaluate information. This connects to the effects spiral (Corollary A.3.1): once an agent's trust calibration degrades, it becomes increasingly vulnerable to further deception, creating a positive-feedback loop in model corruption.

Symmetry with cooperative query actions. The same formal structure — communicative actions with trust-dependent gain operating on another agent's model — encompasses both the cooperative case (teaching, consulting, honest signaling) and the adversarial case (deception, disinformation, strategic ambiguity). What differs is alignment: whether the source's response is optimized to improve or to degrade the receiver's model. The game-theoretic literature on cheap talk, signaling games, and mechanism design addresses when honest communication is incentive-compatible — a question TFT does not attempt to answer but whose relevance it makes formally precise.

Multi-agent and game-theoretic extensions. Appendix F extends the query-action framework to $N$-agent networks, formalizing the communication gain (a multi-source extension of TF-06's uncertainty ratio that adds source-quality and alignment uncertainty terms), and identifies where game-theoretic equilibrium analysis is needed to determine strategic communication incentives. Decision-theoretic calibration of $\lambda$ and the deliberation stopping rule are operationalized in Appendix B, Section B.6.

Domain Instantiations

Domain Exploration: direct probing Exploration: query actions
Kalman + LQR Dual control (rare) — (no external models)
RL ε-greedy, UCB, Thompson Imitation learning, reward shaping from demonstrations
PID Perturbation testing Consulting plant specifications
Boyd's OODA Probing, feints, recon Intelligence gathering, interrogation, liaison with allies
Organism Play, curiosity, foraging Social learning, asking, observing experts
Organization R&D, experiments, pilots Hiring consultants, benchmarking, acquiring companies
Science Experimentation Literature review, peer consultation, conference attendance
Immune Random antibody generation Maternal antibodies, microbiome signaling

Footnotes

  1. Feldbaum, A. A. (1960). "Dual control theory I–IV." Avtomatika i Telemekhanika, 21(9).