Replicator Dynamics Nash makes sense if… Rational Calculating Such as Auctions… Or Oligopolies… But why would Nash matter for our puzzles? Norms/rights/morality are not chosen after careful consideration at the time they influence our behavior We believe we have rights We feel angry when uses service but doesn’t pay We kinda like the mug the experimenter gave us Rather, they are adopted beforehand via prestige-biased imitation reinforcement learning evolution Do these dynamic processes lead to Nash? Yes. We’ll confirm this by modeling prestigebiased learning and evolution in games using the canonical model: the replicator dynamic We consider a large population, , of players Each period, a player is randomly matched with another player and they play a two-player game Each player is assigned a strategy. Players cannot choose their strategies We can think of this in a few ways, e.g.,: Players are “born with” their mother’s strategy (ignore sexual reproduction) Players “imitate” others’ strategies Note that rationality and conscious deliberation don’t enter the picture Suppose there are two strategies, and We start with: Some number, , of players assigned strategy and some number, , of players assigned strategy We denote the proportion of the population playing strategy as , so: The state of the population is given by where , , and Since players interact with another randomly chosen player in the population, a player’s expected payoff is determined by the payoff matrix and the proportion of each strategy in the population For example, consider the coordination game: And the following starting frequencies: Payoff for player who is playing Since depends on and is we write We interpret payoffs as rate of reproduction (fitness) The average fitness, , of a population is the weighted average of the two fitness values. How fast do and grow? Recall First, we need to know how fast Let grows Each individual reproduces at a rate Next we need to know how fast , and there are of them. So: grows. By the same logic: By the quotient rule, and with a little simplification… This is the replicator equation: Growth rate of A Current frequency of strategy Own fitness relative to the average This is the replicator equation: Growth rate of A Current frequency of strategy Own fitness relative to the average Because that’s how many As can reproduce This is our key property. More successful strategies grow faster If: : The proportion of As is non-zero : The fitness of A is above average Then: : A will be increasing in the population We are interested in states where since these are places where the proportions of A and B are stationary We will call such points steady states The steady states are such that Recall the payoffs of our (coordination) game: A B A a b B c d a>c b<d = “asymptotically stable” steady states i.e. steady states s.t. the dynamics point toward it What were the pure Nash equilibria of the coordination game? A B A a, a b, c B c, b d, d Replicator predicts we’ll end up at one of these And the mixed strategy equilibrium is: Replicator teaches us: We end up at Nash (…if we end anywhere) AND not just any Nash (e.g., not mixed Nash in coordination) Let’s generalize this to three strategies: Now… is the number playing is the number playing is the number playing Now… is the proportion playing is the proportion playing is the proportion playing The state of population is where , and , , For example, Consider the Rock-Paper-Scissors Game: R P S R 0 -1 1 P 1 0 -1 S -1 1 0 With starting frequencies: Fitness for player playing is In general, fitness for players with strategy R is: The average fitness, f, of the population is: Replicator is still: Current frequency of strategy Own fitness relative to the average Notice not asymptotically stable It cycles R P S R 0 -1 2 P 2 0 -1 S -1 2 0 Is now asymptotically stable
© Copyright 2026 Paperzz