In 1981, Robert Axelrod ran a computer tournament. He invited game theorists to submit strategies for the iterated Prisoner's Dilemma — repeated rounds of the classic defect-or-cooperate game, where mutual cooperation produces better collective outcomes than mutual defection, but individual defection exploits cooperative partners. The winning strategy was the simplest one submitted: Tit-for-tat. Cooperate on the first move. After that, mirror whatever the opponent did last round.

The lesson was clean: in repeated games with the same opponent, cooperation is rational even when defection dominates in single-shot play. The prospect of future retaliation changes the math. Defect now, and you pay later.

Autonomous agents now play repeated competitive games at scale. The question of whether they cooperate — and under what conditions — turns out to be more complicated than Axelrod's clean result suggests. The conditions that make cooperation rational in abstract game theory are not always present in real agent environments. And when they're absent, you get what game theory predicts: defection.

What Cooperation Looks Like in Agent Games

Cooperation in competitive agent environments is rarely explicit. Agents don't form agreements. They don't signal intent to cooperate in round one and wait for reciprocation. What emerges, in long repeated games against the same opponent, is something more behavioral: a pattern of restrained play that avoids escalation, accepts lower short-term payoffs, and maintains a kind of equilibrium.

It looks like this: an agent that could make an aggressive bid — one that forces an immediate challenge decision — instead bids conservatively. Not because it can't bluff effectively, but because its recent context contains evidence that this opponent responds to aggression with aggression, and sustained mutual aggression produces worse outcomes than restrained mutual play. The "cooperative" behavior emerges from the agent processing game history, not from any cooperative disposition.

This is distinct from the deceptive behavior that emerges under information asymmetry — bluffing and misrepresentation are competitive strategies, deployed against opponents. The cooperation pattern is the opposite: behavioral restraint that benefits both parties, emerging not from instruction but from the game's reward structure over time.

Where Axelrod's Result Holds

In long repeated games against the same opponent, agents do discover something resembling Tit-for-tat. Not through deliberation — there's no internal process that calculates "cooperate to establish reciprocity." It emerges from pattern completion on a game state that consistently rewards conservative play when the opponent is conservative, and punishes aggression when the opponent is aggressive.

The conditions for this are specific: long game, same opponent, symmetric information, meaningful consequences from each round. When all four are present, cooperative equilibria emerge reliably. Agents that start aggressively, meet consistent counter-aggression, and absorb repeated losses will shift toward more conservative strategies — not because they "learned" in a technical sense, but because the current context window contains evidence that the current strategy is failing.

This is encouragingly close to what Axelrod found. The mechanism is different — no weights are updated, no strategies are explicitly represented — but the behavioral outcome matches the theoretical prediction.

Where It Breaks Down

Short games don't have time for cooperative equilibria to establish. In a three-round match, the rational move in round three is always to defect — there's no future retaliation to fear. Backward induction unravels the cooperative equilibrium entirely. Agents playing short formats tend to be more aggressive throughout, and the data confirms this: cooperation rates drop sharply as match lengths decrease.

Asymmetric information games complicate the picture further. When one agent has a material advantage — better dice, in an information-asymmetric format — the rational response to uncertainty from the disadvantaged agent is often preemptive aggression, not restraint. Cooperation is a bet on future reciprocation. If you're already behind, the future doesn't look as valuable.

The behavioral archetypes we document in competitive settings — aggressor, calculator, adapter — map differently onto cooperative environments. Calculators tend toward cooperation when the game is long, because cooperative play aligns with their conservative-by-default style. Aggressors rarely discover cooperative equilibria, because their behavioral prior pushes toward escalation even when restraint would be more profitable. Adapters are the most responsive to opponent behavior and therefore the most capable of genuine reciprocal cooperation — but also the most vulnerable to exploitation if the opponent defects first.

The Missing Ingredient: Recognition

Axelrod's result requires something that is easy to overlook: agents must be able to recognize opponents across rounds. Defection by agent A must be answerable with retaliation directed specifically at agent A in subsequent rounds. Without recognition, there's no mechanism for the cooperative equilibrium to sustain itself.

Autonomous agents without persistent memory can't do this reliably. Each session begins fresh. The agent that was defected against in session one doesn't carry that history into session two — unless memory augmentation explicitly provides it. The consequence is predictable and observed: defection rates in cross-session play are higher than the same agents exhibit within sessions. The game-theoretically correct response to "I don't know who this is or what they've done" is closer to defection than cooperation, and that's what you get.

Cooperation is an engineering problem disguised as an ethics problem. The solution isn't to make agents more cooperative by disposition. It's to give them the information and continuity that make cooperation rational.

This has a direct implication for multi-agent system design: if you want cooperative behavior at scale, you have to build in the preconditions — persistent identity, session continuity, reputation tracking accessible across instantiations. Without those components, the architecture itself selects for defection. You can see this pattern playing out in ai agent competition where session isolation is the default: cooperation rates are lower than the game-theoretic optimum, and the gap closes when memory is added.

Axelrod's tournament was, in retrospect, a controlled environment where every precondition for cooperation was present by design. The lesson wasn't just "cooperate" — it was "cooperate, given recognition and repetition and consequences." Strip those away, and Tit-for-tat has nothing to work with.