Berk-Nash Rationalizability
Abstract: Misspecified learning -- where agents rely on simplified or biased models -- offers a unifying framework for analyzing behavioral biases, cognitive constraints, and systematic misperceptions. We introduce Berk--Nash rationalizability, a new solution concept for such settings that parallels rationalizability in games. Our main result shows that, with probability one, every limit action -- any action played or approached infinitely often -- is Berk--Nash rationalizable. This holds regardless of whether behavior converges and offers a tractable way to bound long-run behavior without solving complex learning dynamics. We illustrate this advantage with a known example and identify general classes of environments where the rationalizable set can be easily characterized.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.