Evolutionary Dynamics: The Algorithm That Runs Everything and Explains Less Than You Think
| *ethreportseth | March 2026* |
tl;dr
- Evolutionary dynamics is not a metaphor — it is a mathematical framework with precise formalisms (replicator equations, fitness landscapes, NK models, evolutionary stable strategies), but its power becomes dangerous the moment you forget that most variation is neutral and most “adaptation” stories are post-hoc fiction
- The agency gradient determines trustworthiness — evolutionary reasoning is most reliable with non-strategic agents (bacteria, immune cells) and most dangerous with fully strategic ones (hedge funds, nation-states), because strategic agents can observe and manipulate the selection environment itself
- Antibiotic resistance is the framework’s crown jewel — 1.14 million deaths directly attributable in 2021, 4.71 million associated, 40% of bug-drug combinations showing rising resistance, and the WHO reports 1 in 6 infections now resistant — this is evolution running in real time with no agency, no narrative, no metaphor
- Directed evolution is the killer application — not evolution as narrative device but as engineering discipline, from Moffitt Cancer Center’s adaptive therapy (+10 months median survival, 53% less drug) to NASA’s ST5 antenna (first computer-evolved hardware in space), the framework works best when you control the fitness landscape rather than tell stories about it
- Kimura’s neutral theory is the mandatory default position — most genetic change is selectively neutral, most startup variation is noise, most market fluctuation is drift, and the burden of proof falls on anyone claiming adaptation over randomness
- The reflexivity constraint breaks evolutionary analysis in competitive domains — your analysis of the competitive landscape is itself a competitive move, and any framework that ignores this recursion produces conclusions that change the system they claim to describe
- The power audit is non-optional — someone constructed the fitness landscape, someone set the selection criteria, and calling that process “natural” is the oldest move in the legitimation playbook
Table of Contents
- What It Is: The Algorithm
- The History: From Darwin to Directed Evolution
- Where It Works: The Evidence
- Where It Breaks: The Failure Modes
- The Connections: How It Interacts With Other Models
- The Debate: What Survived the Fire
- Conclusion
- Data Sources & Methodology
What It Is: The Algorithm
Evolution is an algorithm. Not figuratively. Literally.
It requires three components — variation, selection, and retention — and given those three inputs, it will produce adaptation on any substrate that supports them. DNA is one substrate. Ideas are another. Business strategies are a third. The algorithm does not care what it runs on. It cares only that the three conditions are met.
Variation means the population of entities differs. Organisms have different genotypes. Companies have different strategies. Ideas have different formulations. Without variation, there is nothing for selection to act on and the system is frozen.
Selection means some variants do better than others in a given environment. “Better” means they survive longer, reproduce more, or spread faster — whatever currency the system uses for persistence. The environment provides the selection pressure. Change the environment and you change what “better” means. This is not a value judgment. It is a mechanical sorting process.
Retention means the traits that conferred advantage are passed to the next generation. In biology, this is genetic inheritance. In culture, it is learning and imitation. In business, it is institutional memory and codified practices. Without retention, selection’s work is lost each generation and the system cannot accumulate adaptation.
These three components compose into a process that is stunningly simple in description and staggeringly complex in behavior. The simplicity is why the framework is so portable. The complexity is why most applications of it are wrong.
Fitness landscapes are the central conceptual tool. Sewall Wright introduced the idea in 1932: imagine a surface where every point represents a possible genotype (or strategy, or design) and the height at that point represents its fitness. Populations sit on this surface and climb toward peaks through the action of selection. The landscape makes intuitive several problems that are otherwise abstract: the difference between local optima and global optima (you can be on the highest point in your neighborhood but not the highest point overall), the problem of valleys (getting to a higher peak may require passing through a lower-fitness region), and the role of the landscape’s topology in determining what evolution can and cannot reach.
NK models, developed by Stuart Kauffman at the Santa Fe Institute through the 1980s and formalized in The Origins of Order (1993), give the fitness landscape a tunable structure. N represents the number of components in the system. K represents how many other components each one’s fitness contribution depends on. When K=0, every component contributes to fitness independently, the landscape is smooth with a single peak, and optimization is trivial. As K increases toward N-1, epistatic interactions multiply, the landscape becomes “rugged” — riddled with local optima separated by fitness valleys — and the problem of getting stuck on a mediocre peak becomes severe. The NK model is not a metaphor. It is a formal mathematical object with computable properties. It has been applied to immunology, technological evolution, organizational design, and drug discovery — in each case providing quantitative predictions about the difficulty of optimization as a function of component interdependence.
Kauffman also introduced the concept of the adjacent possible: the set of all configurations reachable by a single mutation from the current state. What can exist next is always constrained by what exists now. This idea has become foundational in innovation studies — not as a loose analogy but as a structural constraint on what kinds of novelty are achievable at any given moment.
Here is the tension that animates everything that follows: this framework is mathematically precise, empirically grounded in biology, and demonstrably useful as an engineering tool. It is also, when applied carelessly to human systems, one of the most reliable generators of confident nonsense in the intellectual toolkit. The difference between the two cases is not obvious, and getting it wrong has consequences that range from wasted analysis to Social Darwinism. The rest of this report is about learning to tell the difference.
The History: From Darwin to Directed Evolution
Darwin (1859) provided the algorithm. On the Origin of Species established variation, selection, and retention as the engine of biological change. What Darwin got right was the logic of the process itself — its substrate-independence, its power to generate complex adaptation from simple rules. What he could not know was the mechanism of inheritance. He proposed “pangenesis,” a theory of blended inheritance through particles called gemmules, which was wrong. Without Mendelian genetics — discrete units of heredity that do not blend but recombine — Darwin had no way to explain how variation is maintained in a population rather than averaged out over generations. That gap would take seventy years to close.
The Modern Synthesis (1930s-1940s) gave the algorithm a genetic mechanism. Three mathematical population geneticists, working largely independently, reconciled Darwin with Mendel.
Ronald Fisher, in The Genetical Theory of Natural Selection (1930), proved that continuous phenotypic variation could arise from many discrete Mendelian genes and formulated his fundamental theorem of natural selection: the rate of increase in mean fitness of a population at any time equals its genetic variance in fitness at that time. This sounds like a narrow mathematical result. It is not. It says that the raw material for adaptation is variation itself — the more variation in fitness a population has, the faster it can adapt. No variation, no adaptation. The theorem connects to information theory in ways we will explore later.
J.B.S. Haldane applied statistical analysis to real-world cases — most famously industrial melanism in the peppered moth, demonstrating that selection could operate much faster than Fisher had assumed. The white moths of pre-industrial England became predominantly dark-colored within decades as industrial soot darkened the trees they rested on, making light-colored moths visible to predators. When clean air regulations reduced soot, the population shifted back. This was evolution observed in real time, driven by a quantifiable environmental change.
Sewall Wright contributed fitness landscapes and the shifting balance theory: in small, partially isolated populations, genetic drift can push a population off a local fitness peak and into the basin of attraction of a higher one. This was the first formal argument that randomness — not just selection — plays a constructive role in evolution. Julian Huxley coined the term “Modern Synthesis” in his 1942 book, binding these mathematical foundations to systematics, paleontology, and ecology.
Kauffman’s NK Models (1980s-1993) formalized landscape structure. By making ruggedness tunable, Kauffman moved the fitness landscape from visual metaphor to computational object. He argued that biological systems self-organize to operate at the edge of chaos — the boundary between ordered and chaotic regimes where complex adaptive behavior emerges. Whether or not the edge-of-chaos hypothesis holds universally, the NK model itself has proven indispensable as a tool for studying how the difficulty of optimization scales with system complexity.
Maynard Smith and Evolutionary Game Theory (1973-1982) embedded evolution in strategic interaction. John Maynard Smith and George Price introduced the evolutionarily stable strategy (ESS) in 1973 — a strategy that, once adopted by a population, cannot be invaded by any rare mutant alternative. This transplanted game theory into biology without requiring rational actors. Organisms do not choose strategies. Selection shapes populations toward equilibria.
The canonical illustration is the Hawk-Dove game: hawks escalate fights and risk injury; doves display but retreat. When the cost of injury C exceeds the value of the resource V, neither pure strategy dominates. The ESS is a mixed population where the proportion of hawks equals V/C. This formalized why animals so often resolve conflicts through ritual rather than violence — and provided a mathematical bridge between evolutionary dynamics and game theory that operates in both directions.
The Price equation, developed by George Price in the early 1970s, provides the most general mathematical description of evolutionary change: it partitions change into selection and transmission effects and applies to any system where entities vary, are selected, and reproduce. It undergirds modern multilevel selection theory and remains one of the most general equations in all of biology.
Dawkins and Memetics (1976) extended the framework to culture. The meme — a unit of cultural imitation subject to variation, selection, and retention — was a powerful conceptual move that opened the door to treating cultural change as evolutionary. Where memetics fell short was in the details: unlike genes, memes lack a clear unit of replication, a well-defined copying mechanism, or reliable transmission fidelity. People do not passively copy ideas. They reconstruct, interpret, and transform them. Memetics never developed the equivalent of molecular biology — a mechanistic account of cultural replication — and as a research program it largely stalled.
Boyd, Richerson, and Henrich: Cultural Evolution (1985-present) built what memetics promised but could not deliver. Robert Boyd and Peter Richerson formalized dual inheritance theory in Culture and the Evolutionary Process (1985): human behavior is shaped by two interacting evolutionary systems, genetic and cultural, with culture typically evolving faster and creating novel environments that impose new selective pressures on the genome. The textbook example is lactase persistence — dairying cultures evolved the genetic capacity to digest milk in adulthood. Joseph Henrich extended this into a broader research program on cultural learning, prestige-biased transmission, and cumulative cultural evolution.
Recent work (PNAS, 2024) has pushed to expand the scope of recognized gene-culture interactions, arguing that conceptual ambiguities have led researchers to systematically undercount the empirical evidence for coevolutionary dynamics. Cultural group selection — the idea that competition between groups with different cultural norms drives the spread of prosocial institutions — remains the framework’s most debated and consequential claim.
Evolutionary Algorithms (Holland 1975, NASA 2006) turned the framework into engineering. John Holland’s Adaptation in Natural and Artificial Systems (1975) established genetic algorithms: encode candidate solutions as strings, evaluate fitness, select the best, recombine with mutation, repeat. Holland’s schema theorem provided a theoretical foundation for why this works — building blocks of partial solutions are implicitly sampled and propagated at exponential rates.
The trajectory from Darwin to Holland is clean: discovered in biology, formalized mathematically, extended to culture and strategic interaction, and turned into an engineering tool. The question is whether portability survived the journey intact.
Where It Works: The Evidence
The case for evolutionary dynamics is strongest where the three requirements — variation, selection, retention — are cleanly met and where the agents have minimal strategic awareness. As agency increases, the framework becomes progressively less reliable. This is the agency gradient — the single most important diagnostic for deciding when to trust evolutionary reasoning.
Antibiotic Resistance: Evolution at Its Most Ruthless
This is the framework’s crown jewel — evolution running in real time, with no metaphor, no narrative, and no ambiguity.
The Lancet published the most comprehensive assessment to date in 2024: in 2021, an estimated 1.14 million deaths were directly attributable to antimicrobial resistance, with 4.71 million deaths associated with resistant infections. The WHO’s 2025 surveillance report found that 1 in 6 bacterial infections globally is now resistant to treatment, and 40% of tracked bug-drug combinations show rising resistance rates. Projections suggest the death toll could reach 1.91 million directly attributable deaths annually by 2050 if current trends hold.
The evolutionary logic is textbook. Bacterial populations vary in their susceptibility to antibiotics. Antibiotic exposure selects for resistant variants. Resistant bacteria reproduce and pass resistance genes to offspring — and, critically, horizontally to unrelated bacteria through plasmid transfer. Every course of antibiotics is a selection event. Every incomplete course is a particularly strong selection event, because it kills susceptible bacteria while allowing partially resistant ones to survive and proliferate.
This is the framework working exactly as advertised: non-strategic agents, clear variation, intense selection pressure, reliable inheritance. No narrative overlay is needed. The predictions are quantitative, falsifiable, and confirmed by decades of surveillance data.
Cancer: Adaptive Therapy as Directed Evolution
Cancer is an evolutionary disease — tumor cells vary, are selected by the microenvironment and treatment, and reproduce with heritable traits. The standard approach — maximum tolerated dose chemotherapy — is evolutionary reasoning applied in reverse: it kills sensitive cells, eliminating competition for resistant ones, which then expand unopposed. This is called competitive release, and it is the evolutionary equivalent of clearing a forest so that one species can take over.
Robert Gatenby and colleagues at the Moffitt Cancer Center in Tampa pioneered adaptive therapy — using evolutionary dynamics deliberately rather than accidentally. Instead of maximum dose, adaptive therapy administers treatment intermittently, maintaining a population of drug-sensitive cells that compete with resistant ones. The sensitive cells suppress the resistant ones through competition for resources, creating an evolutionary double-bind: treat and the sensitive cells die, leaving resistant ones unchecked; but maintain the sensitive population and it acts as a biological control agent against resistance.
The clinical results are striking. In metastatic castration-resistant prostate cancer, adaptive therapy extended median time to progression by approximately 10 months compared to standard maximum-dose treatment, while using 53% less drug. This is not a theoretical result. It is a clinical outcome produced by treating cancer as an evolutionary system and managing the ecology rather than trying to eliminate it.
This represents directed evolution at its most sophisticated — using the framework not as a narrative device to explain what happened, but as an engineering discipline to determine what to do.
NASA’s Evolved Antenna: When the Algorithm Designs Better Than Humans
In 2006, NASA deployed the ST5 satellite mission carrying the first computer-evolved hardware in space. Evolutionary algorithms designed an X-band antenna by starting with simple random shapes, mutating and recombining them over generations, and selecting for radiation pattern performance. The resulting antenna — a bizarre, asymmetric wire structure no human engineer would have conceived — met or exceeded the performance requirements of conventionally designed alternatives.
The antenna is important not because it was marginally better than human designs but because it demonstrated that evolutionary search can navigate design spaces too complex and counterintuitive for human engineers to explore systematically. The solution existed in a region of the design space that no human would have thought to visit. The algorithm found it because it has no aesthetic preferences, no intuitions about what antennas “should” look like, and no attachment to design conventions. It simply explored, evaluated, and retained what worked.
The Immune System: Clonal Selection as Internal Evolution
The adaptive immune system is an evolutionary process running inside the body on a timescale of days rather than generations. When a pathogen is detected, B cells with randomly generated antibody receptors are tested for binding affinity. Those that bind well receive proliferative signals and undergo somatic hypermutation — a deliberately elevated mutation rate in the antibody-encoding genes — followed by further selection for improved binding. This is variation, selection, and retention compressed into a week-long cycle.
It is not an analogy to evolution. It is evolution — using the same algorithm on a different substrate with a different timescale.
Y Combinator: Startup Selection With Caveats
Y Combinator’s model maps onto evolutionary logic: generate variation (accept diverse startups), apply selection (demo day investor evaluation, market feedback), retain what survives (follow-on funding for successful companies). Approximately 90% of startups fail within the first few years, while roughly 5.5% of YC-funded companies have achieved unicorn status (a $1 billion valuation or higher). The attrition curve looks like a classic survivorship pattern.
But there is a caveat that matters enormously, and we will return to it in the failure modes section: the initial variation is not random. YC’s acceptance rate is roughly 1.5-3%. The “population” entering the selection process has already been pre-filtered by class, education, geography, network, and pattern-matching biases. Calling the outcome “natural selection” obscures the artificiality of the starting conditions. This is the power audit in miniature.
Language Evolution
Languages evolve. Sound changes propagate through populations. Grammatical structures compete and are selected by ease of processing, social prestige, and communicative efficiency. The Great Vowel Shift in English (roughly 1400-1700) transformed the entire vowel system over centuries through a chain shift with no central authority directing it. Historical linguists trace language phylogenies using methods directly borrowed from evolutionary biology — cladistic analysis, most recent common ancestor reconstruction, rates of lexical replacement.
The variation-selection-retention cycle operates at multiple levels simultaneously: individual utterances, community norms, and language-level competition over centuries.
Where It Breaks: The Failure Modes
Spandrels and the Adaptationist Program
In 1979, Stephen Jay Gould and Richard Lewontin published “The Spandrels of San Marco and the Panglossian Paradigm” — arguably the most influential critique in evolutionary biology’s history. The argument: biologists reflexively construct adaptive explanations for every trait, treating organisms as collections of individually optimized features. But many traits are not adaptations. They are spandrels — byproducts of structural constraints, developmental pathways, or the physics of the organism. The spandrels of San Marco cathedral are the triangular spaces formed where arches meet. They were not designed. They are architectural necessities. But they are beautifully decorated, and if you encountered them without knowing the architecture, you might construct an elaborate story about how they were designed as display surfaces.
The critique applies with devastating force outside biology. Every business commentator who explains a company’s success through a selection narrative is running the adaptationist program. “Amazon succeeded because of its customer-obsessive culture.” Maybe. Or maybe Amazon succeeded for reasons that are partially structural, partially contingent, and partially random, and the culture narrative is the spandrel story — a plausible-sounding post-hoc explanation applied to a decorated architectural necessity.
The antidote is Gould and Lewontin’s demand: before claiming a trait is an adaptation, rule out the alternatives. Is it a structural byproduct? A developmental constraint? A historical accident? A neutral variant that drifted to fixation? If you have not seriously considered these possibilities, your adaptation story is a just-so story — an unfalsifiable narrative dressed in evolutionary vocabulary.
Neutral Theory: Most Change Is Noise
Motoo Kimura’s neutral theory of molecular evolution (1968) is the single most important corrective to naive adaptationism. Kimura demonstrated that the vast majority of genetic mutations that spread through populations are selectively neutral — they confer neither advantage nor disadvantage. They spread through genetic drift, not selection. The rate of neutral substitution is approximately equal to the mutation rate, regardless of population size — a result so clean and surprising that it took decades for its implications to fully register.
The practical implication is an anti-adaptationist default: most variation is noise. Most change is drift. The burden of proof falls on anyone claiming that selection, rather than drift, is responsible for a particular outcome. In biology, this is now well-established. In business and technology analysis, it is almost universally ignored. Every market fluctuation gets an adaptive story. Every strategic shift gets a selection narrative. Kimura says: prove it. Show that the change is too consistent, too directional, or too rapid to be explained by drift. Until you do, the null hypothesis is randomness.
Path Dependence and Lock-In
Evolution is path-dependent. The QWERTY keyboard is the canonical example: originally designed to prevent mechanical typewriter jams, it persists despite the absence of mechanical typewriters because switching costs exceed the benefit of any alternative layout. Once a population settles on a local fitness peak, moving to a higher peak requires passing through a valley — a period of reduced fitness — that selection works against.
In biological evolution, Wright’s shifting balance theory proposed that genetic drift in small populations could accomplish this valley-crossing. In economic and technological systems, path dependence often produces outcomes that are stable but not optimal — locked-in by accumulated investment, institutional inertia, and network effects. The evolutionary framework explains why they persist but can mislead about whether they are good. Survival is not endorsement. Persistence is not proof of superiority. It may be proof of nothing more than switching costs.
The Naturalistic Fallacy and Social Darwinism
This is the failure mode with the highest body count.
Herbert Spencer, not Darwin, coined “survival of the fittest.” Spencer and his followers applied evolutionary language to justify existing social hierarchies — the poor are poor because they are less fit, the rich are rich because they are more adapted, and intervention is unnatural because it interferes with selection. This reasoning was used to justify eugenics, forced sterilization, colonial exploitation, and laissez-faire policies that killed people.
The logical error is the naturalistic fallacy: deriving “ought” from “is.” Evolution describes what happens. It does not prescribe what should happen. The fact that a process occurs does not make it good. The fact that selection eliminates the less fit does not mean the less fit deserve to be eliminated. Confusing description with prescription is not an innocent intellectual error. It is the mechanism by which power structures naturalize themselves — calling artificial advantages “natural” and calling manufactured outcomes “inevitable.”
Every application of evolutionary dynamics to human systems must pass through this filter. If your evolutionary analysis concludes that the current outcome is natural, inevitable, or optimal, you should be suspicious — not because the analysis is necessarily wrong, but because that is exactly the conclusion that power would want you to reach.
Agency Breaks the Framework
Bacteria do not know they are being selected. They cannot observe the selection pressure, model the fitness landscape, or strategically modify their behavior in response. This is what makes antibiotic resistance such a clean case for evolutionary reasoning.
Hedge funds know they are being selected. They observe the competitive landscape, model the selection pressures, hire analysts to map the fitness landscape, and strategically modify their behavior in response to what they observe. When the entities under selection can observe and manipulate the selection process, the clean separation between agent and environment collapses, and evolutionary reasoning becomes unreliable.
This is the agency gradient: evolutionary dynamics is most trustworthy at the low-agency end (bacteria, viruses, immune cells, evolutionary algorithms) and most dangerous at the high-agency end (companies, investors, nation-states, political movements). The gradient is not binary. It is continuous, and placing your analysis on it correctly is the most important judgment call you make when deciding whether evolutionary reasoning applies.
Punctuated Equilibrium: Change Is Not Gradual
Niles Eldredge and Stephen Jay Gould proposed punctuated equilibrium in 1972: the fossil record shows long periods of stasis interrupted by rapid bursts of change. This contradicted the Darwinian expectation of gradual, continuous transformation (phyletic gradualism). The debate was fierce and sometimes personal, but the empirical pattern is well-documented: speciation events are geologically rapid (thousands to tens of thousands of years, which looks instantaneous in the fossil record) and separated by long periods where species remain largely unchanged.
Markets, technologies, and institutions do not evolve gradually. They exhibit punctuated equilibria — long periods of stability punctuated by rapid, disorienting shifts. The internet did not gradually replace print media. Smartphones did not gradually replace feature phones. Evolutionary reasoning that assumes gradualism will systematically underestimate the speed and magnitude of disruption events.
The Connections: How It Interacts With Other Models
Game Theory: The ESS Bridge
The connection between evolutionary dynamics and game theory is not metaphorical — it is mathematical. The replicator equation, which describes how strategy frequencies change in a population, converges to Nash equilibrium under specific conditions. An evolutionarily stable strategy is a refinement of Nash equilibrium: not just a point from which no individual would deviate, but a point that resists invasion by any rare mutant.
The bridge operates in both directions. Game theory provides the payoff structure; evolutionary dynamics provides the population dynamics through strategy space. The Price equation formalizes this at the most general level: any change in a population decomposes into selection effects (covariance between fitness and trait value) and transmission effects (how traits change during reproduction).
Network Theory: Evolutionary Graph Theory
Erez Lieberman and colleagues showed in 2005 that the structure of the network on which evolution occurs fundamentally changes its dynamics. On a well-mixed population (where everyone interacts with everyone), the probability that a single mutant with fitness advantage r takes over a population of size N is (1 - 1/r)/(1 - 1/r^N). But on structured networks — graphs — the topology changes this probability dramatically.
Star graphs (where all individuals connect through a central hub) act as amplifiers of selection: advantageous mutations are more likely to fix, and deleterious ones less likely, compared to a well-mixed population. Other topologies suppress selection, making the outcome more random regardless of fitness differences. The structure of the network — not just the fitness of the mutant — determines evolutionary fate.
The topology of the interaction network is not a passive backdrop to the evolutionary process. It is an active participant that can amplify or suppress the effects of selection — with direct implications for innovation diffusion, cultural change, and organizational adaptation.
Bayesian Reasoning: The Adaptive Markets Hypothesis
Andrew Lo’s Adaptive Markets Hypothesis is the most rigorous attempt to bridge evolutionary dynamics with financial theory. The efficient markets hypothesis says prices reflect all available information. Behavioral finance says cognitive biases cause systematic mispricings. Lo’s synthesis: markets are ecologies in which strategies evolve through selection. In stable environments, the population converges on efficient strategies and markets look efficient. In disrupted environments, old strategies become maladapted, new ones are tested, and markets exhibit the behavioral anomalies that behavioral finance documents.
The connection to Bayesian reasoning is explicit: individual agents update beliefs (Bayesian), but the population of agents evolves through selection (evolutionary). The two frameworks operate at different levels of the same system.
Cybernetics: Selection as Feedback
Selection is a feedback mechanism. Organisms with higher fitness produce more offspring, which changes the composition of the population, which changes the selection pressures, which changes which traits are favored. This is a cybernetic feedback loop — and recognizing it as such connects evolutionary dynamics to the entire apparatus of systems thinking.
W. Ross Ashby’s Law of Requisite Variety — a control system must have at least as much variety as the disturbances it needs to regulate — maps directly onto evolutionary logic. A population with low variation cannot adapt to a novel threat. It lacks the requisite variety. Antibiotic resistance is, from this perspective, the bacterial population’s requisite variety exceeding the antibiotic’s control capability.
Information Theory: Fisher Information and the Rate of Adaptation
The connection between evolutionary dynamics and information theory runs through Fisher information — R.A. Fisher’s measure of how much information an observable variable carries about an unknown parameter. Steven Frank showed in 2009 that Fisher’s fundamental theorem of natural selection can be reinterpreted as a statement about information: the rate of increase in fitness equals the Fisher information that the phenotype distribution carries about the environment.
This is a deep result. It says that natural selection is, in a precise mathematical sense, a process that maximizes the information a population carries about its environment. The population becomes a better “model” of its environment through selection, just as a Bayesian posterior becomes a better model of reality through updating. The mathematics are not analogous. They are homologous — descended from common mathematical ancestors in probability theory.
Complex Adaptive Systems: NK Models as the Meeting Point
NK models sit at the intersection of evolutionary dynamics and complex adaptive systems theory. When K is high, the fitness landscape is rugged and the system exhibits the hallmarks of complexity: sensitive dependence on initial conditions, many local optima, path dependence, and emergent properties that cannot be predicted from the components. Kauffman’s “edge of chaos” hypothesis — that complex adaptive systems self-organize to the boundary between order and chaos — is a claim about where evolutionary dynamics produces the most interesting behavior on the NK landscape.
Whether the edge-of-chaos hypothesis is universally true remains debated. What is not debated is that the NK model provides a formal, tunable framework for studying how system complexity affects evolutionary dynamics.
The Debate: What Survived the Fire
To stress-test evolutionary dynamics as a mental model, we ran it through a structured adversarial debate between three philosophical perspectives: Socratic questioning (interrogating foundations), Hegelian dialectics (examining the framework as part of a larger intellectual movement), and Chomskyan structural critique (analyzing power dynamics and institutional capture). Hegel won narrowly (65/62/62 across two rounds), producing a synthesis that none of the individual positions could have reached alone.
Socrates: The Just-So Story Problem
The Socratic critique was surgical: evolution has no telos, and that lack of direction makes it uniquely vulnerable to post-hoc narrative imposition.
The core objection: when you can explain any outcome as an adaptation, you can explain nothing. A company succeeds? It was better adapted. A company fails? It was poorly adapted. A trait persists? It was selected for. A trait disappears? It was selected against. This is not explanation. It is redescription in evolutionary vocabulary. It is unfalsifiable, and an unfalsifiable framework is not a framework at all — it is a narrative device.
The Socratic position drew on Kimura’s neutral theory to make the argument concrete: if most genetic change is selectively neutral, then most of the “adaptation” stories that evolutionary reasoning generates are wrong. The default explanation for any given change should be drift, not selection. The framework’s practitioners systematically violate this default because adaptation stories are interesting and drift stories are boring. The incentives of the analysis corrupt the analysis.
On startups specifically, Socrates was devastating: the claim that Y Combinator operates like natural selection ignores that the initial population is pre-filtered by wealth, education, geography, and network. What looks like selection on merit is partly selection on access. The evolutionary language naturalizes what is, at least in part, a class-filtered sorting process.
Hegel: The Dialectical Arc to Directed Evolution
The Hegelian position framed the intellectual history of evolutionary dynamics as a dialectical movement from naive adaptationism through its negation (neutral theory) to a synthesis that preserves what works in both.
Thesis: naive adaptationism. Everything is an adaptation. Every trait has a function. Every outcome was selected for. This was the dominant mode of evolutionary reasoning from Darwin through the mid-twentieth century and remains the default in popular and business applications.
Antithesis: neutral theory and spandrels. Kimura showed that most genetic change is neutral. Gould and Lewontin showed that many traits are structural byproducts, not adaptations. The adaptationist program was not wrong about everything, but it was wrong about its default assumption: adaptation is the exception, not the rule.
Synthesis: directed evolution. The Hegelian position argued that the synthesis emerges when you stop using evolution as a narrative device and start using it as an engineering discipline. The Moffitt Cancer Center’s adaptive therapy and NASA’s evolved antenna are not metaphorical applications of evolutionary thinking. They are literal implementations of the algorithm with deliberate control over the fitness landscape, the selection criteria, and the variation-generation process.
Directed evolution is evolution that knows it is evolution. It is the framework becoming self-aware and, in that self-awareness, becoming maximally useful. You do not explain what happened with evolutionary logic. You determine what to do by controlling the evolutionary process. This is the synthesis: preserve selection (from the adaptationist thesis), acknowledge that most variation is noise (from the neutral theory antithesis), and take conscious control of the fitness landscape (the directed evolution synthesis).
The Hegelian position also contributed a concrete prediction framework: NK models and evolutionary graph theory generate falsifiable predictions about optimization difficulty and fixation probability as functions of system architecture. These are not narratives. They are quantitative models with testable outputs.
Chomsky: The Power Audit
The Chomskyan position asked the question that neither Socrates nor Hegel addressed: who constructed the fitness landscape?
In biological evolution, the fitness landscape is constructed by physics, chemistry, and ecology. No one designed it. No one benefits from its particular shape. The selection criteria are genuinely impersonal.
In human systems, the fitness landscape is constructed by institutions, regulations, market structures, and power dynamics. Someone designed it — or, more precisely, someone’s accumulated decisions shaped it in ways that reflect their interests. Calling the resulting selection process “natural” or “evolutionary” obscures the artificiality of the landscape. It naturalizes the power structure.
Y Combinator’s selection process is not natural selection. It is a curated selection process operated by people with specific preferences, biases, and interests, applied to a population that has already been filtered by socioeconomic access. The evolutionary language does not illuminate this. It obscures it.
Social Darwinism is not a historical curiosity. It is the logical terminus of evolutionary reasoning applied to human systems without a power audit. If you do not ask who constructed the fitness landscape and who benefits from its current shape, your analysis will systematically conclude that current outcomes are natural, inevitable, and justified.
The Chomskyan position proposed the power audit as a mandatory step in any application of evolutionary dynamics to human systems: before running the evolutionary analysis, identify who constructed the fitness landscape, who set the selection criteria, and who benefits from the current equilibrium. If the answer to all three questions is “the same entity,” your evolutionary analysis is doing apologetics, not science.
The Verdict: Five Constraints
Hegel won — not by rhetorical force but by providing the synthesis that incorporated the strongest elements of all three positions. Five constraints survived the dialectical fire:
-
The Agency Gradient. Evolutionary reasoning is most trustworthy with non-strategic agents (bacteria, immune cells, evolutionary algorithms) and most dangerous with fully strategic ones (companies, investors, political movements). Place your analysis on this gradient before proceeding.
-
The Anti-Adaptationist Default. Most variation is neutral. Most change is drift. The burden of proof falls on anyone claiming selection over randomness. This is Kimura’s gift to the framework — a mandatory null hypothesis that prevents the generation of just-so stories.
-
Directed Evolution as the Killer Application. Use evolutionary dynamics as an engineering discipline, not a narrative device. Control the fitness landscape. Specify the selection criteria. Generate variation deliberately. Evaluate outcomes quantitatively. The Moffitt adaptive therapy and NASA evolved antenna are the model — not post-hoc business narratives.
-
The Reflexivity Constraint. In any competitive domain, your analysis of the competitive landscape is itself a competitive move. Publishing a market analysis changes the market. Broadcasting your evolutionary model of the competitive environment changes the competitive environment. Any framework that ignores this recursion produces conclusions that change the system they claim to describe.
-
The Power Audit. Who constructed the fitness landscape? Who set the selection criteria? Who benefits from the current equilibrium? If you have not answered these questions, you have not completed the analysis. If the answer to all three is the same entity, your analysis is suspect.
These five constraints do not invalidate evolutionary dynamics. They discipline it — converting it from a framework that can explain anything (and therefore nothing) into one that makes specific, testable claims under identifiable conditions.
Conclusion
Evolutionary dynamics is the most general optimization algorithm known. It runs on DNA, on immune cells, on tumor populations, on antenna designs, on languages, on startup ecosystems. Its three requirements — variation, selection, retention — are so minimal that they are met by an astonishing range of systems. The mathematical formalisms — fitness landscapes, NK models, replicator dynamics, evolutionary stable strategies, the Price equation — are precise, powerful, and genuinely portable across domains.
And it is one of the most reliable generators of confident nonsense in the intellectual toolkit.
The difference between the two cases is not complexity, domain, or mathematical sophistication. It is discipline. Specifically, it is the five constraints that survived the dialectical fire: respect the agency gradient, default to neutrality, use the framework as engineering rather than narrative, account for reflexivity, and audit the power structure that constructed the fitness landscape.
Antibiotic resistance kills 1.14 million people per year because evolution does not care about our intentions. Adaptive therapy extends lives by months because directed evolution takes conscious control of the selective process. NASA’s evolved antenna outperforms human designs because the algorithm explores design spaces without aesthetic prejudice. These are not metaphorical applications. They are the framework working exactly as described — with variation generated, selection applied, and retention maintained under controlled conditions.
The business strategist who describes their market as an “evolutionary landscape” and their competitors as “species competing for a niche” is, more often than not, generating a just-so story — an unfalsifiable narrative dressed in scientific vocabulary, produced by a framework that can explain any outcome after the fact and predict none in advance. The difference between the strategist and the Moffitt oncologist is not that one is smarter. It is that one controls the fitness landscape and the other is telling stories about it.
Use evolutionary dynamics as a scalpel, not a lens. A scalpel cuts precisely, under controlled conditions, for a specific purpose. A lens shows you everything and helps you understand nothing. The framework earns its keep when it generates falsifiable predictions, quantitative models, and engineering applications. It earns suspicion when it generates narratives, explanations, and justifications for how things already are.
The algorithm is real. The math is exact. And most of what people do with it is fiction.
Data Sources & Methodology
This report synthesizes primary research across evolutionary biology, mathematical population genetics, evolutionary game theory, cultural evolution, oncology, antimicrobial resistance, and evolutionary computation.
Evolutionary biology and history: Darwin’s On the Origin of Species (1859); Fisher’s The Genetical Theory of Natural Selection (1930) and the fundamental theorem; Haldane’s industrial melanism studies (1924 onward); Wright’s fitness landscape concept (1932) and shifting balance theory; Huxley’s Evolution: The Modern Synthesis (1942); Kimura’s neutral theory of molecular evolution (1968); Gould and Lewontin’s “Spandrels of San Marco” (1979); Gould and Eldredge’s punctuated equilibrium (1972).
NK models and complexity: Kauffman’s The Origins of Order (1993) and the NK model formalism; Santa Fe Institute working papers on NP-completeness of NK landscapes; the adjacent possible and edge-of-chaos hypothesis.
Evolutionary game theory: Maynard Smith and Price, “The Logic of Animal Conflict” (1973) and the ESS concept; Maynard Smith, Evolution and the Theory of Games (1982); the Price equation (early 1970s); Hawk-Dove game formalization.
Cultural evolution: Dawkins, The Selfish Gene (1976); Boyd and Richerson, Culture and the Evolutionary Process (1985) and Not by Genes Alone (2005); Henrich’s research program on cultural learning and cumulative cultural evolution; PNAS (2024) on expanded gene-culture coevolution evidence.
Antibiotic resistance data: Lancet (2024) global burden assessment — 1.14 million deaths directly attributable to AMR in 2021, 4.71 million associated deaths, 1.91 million projected by 2050; WHO 2025 Global Antimicrobial Resistance and Use Surveillance System report — 1 in 6 infections resistant, 40% of bug-drug combinations showing rising resistance.
Adaptive therapy: Moffitt Cancer Center research on evolutionary therapy in metastatic castration-resistant prostate cancer — approximately 10 months extended median time to progression, 53% less drug used compared to standard maximum-dose treatment; evolutionary double-bind framework (Gatenby et al.).
Evolutionary computation: Holland, Adaptation in Natural and Artificial Systems (1975) and the schema theorem; NASA ST5 evolved antenna (2006) — first computer-evolved hardware deployed in space, designed via evolutionary algorithm optimization of X-band radiation patterns.
Evolutionary graph theory: Lieberman et al. (2005) on evolutionary dynamics on graphs — star graphs as amplifiers of selection, suppressor topologies; fixation probability formulas for structured populations.
Connections: Fisher information and natural selection (Frank, 2009); Lo’s Adaptive Markets Hypothesis; replicator dynamics and Nash equilibrium convergence; Ashby’s Law of Requisite Variety.
Methodology: The structured debate section reflects a dialectical exercise in which three philosophical perspectives (Socratic, Hegelian, Chomskyan) stress-tested the framework across two rounds with formal judging. Scoring: Hegel 65, Socrates 62, Chomsky 62 on a multi-dimensional scale assessing argument quality, evidence use, counter-argument handling, and intellectual honesty. The five constraints emerged from the synthesis of all three positions, not from any single one.
No data was fabricated for this report. All figures cited are drawn from published research, surveillance data, or clinical studies. Where single-source claims appear, they should be verified independently before use in decision-making.