Econometrica: Jul 1999, Volume 67, Issue 4
Experience‐weighted Attraction Learning in Normal Form Games
Colin Camerer, Teck Hua HoIn ‘experience‐weighted attraction’ (EWA) learning, strategies have attractions that reflect initial predispositions, are updated based on payoff experience, and determine choice probabilities according to some rule (e.g., logit). A key feature is a parameter that weights the strength of hypothetical reinforcement of strategies that were chosen according to the payoff they would have yielded, relative to reinforcement of chosen strategies according to received payoffs. The other key features are two discount rates, and , which separately discount previous attractions, and an experience weight. EWA includes reinforcement learning and weighted fictitious play (belief learning) as special cases, and hybridizes their key elements. When = 0 and = 0, cumulative choice reinforcement results. When = 1 and =, levels of reinforcement of strategies are exactly the same as expected payoffs given weighted fictitious play beliefs. Using three sets of experimental data, parameter estimates of the model were calibrated on part of the data and used to predict a holdout sample. Estimates of are generally around .50, around .8 − 1, and varies from 0 to . Reinforcement and belief‐learning special cases are generally rejected in favor of EWA, though belief models do better in some constant‐sum games. EWA is able to combine the best features of previous approaches, allowing attractions to begin and grow flexibly as choice reinforcement does, but reinforcing unchosen strategies substantially as belief‐based models implicitly do.
Log In To View Full Content