Fragility of Asymptotic Agreement Under Bayesian Learning
Massachusetts Institute of Technology (MIT) - Department of Economics; Centre for Economic Policy Research (CEPR); National Bureau of Economic Research (NBER)
Massachusetts Institute of Technology (MIT) - Department of Economics; New Economic School
Massachusetts Institute of Technology (MIT) - Department of Economics
March 15, 2008
MIT Department of Economics Working Paper No. 08-09
Under the assumption that individuals know the conditional distributions of signals given the payoff-relevant parameters, existing results conclude that as individuals observe infinitely many signals, their beliefs about the parameters will eventually merge. We first show that these results are fragile when individuals are uncertain about the signal distributions: given any such model, a vanishingly small individual uncertainty about the signal distributions can lead to a substantial (non-vanishing) amount of differences between the asymptotic beliefs. We then characterize the conditions under which a small amount of uncertainty leads only to a small amount of asymptotic disagreement. According to our characterization, this is the case if the uncertainty about the signal distributions is generated by a family with "rapidly-varying tails" (such as the normal or the exponential distributions). However, when this family has "regularly-varying tails" (such as the Pareto, the log-normal, and the t-distributions), a small amount of uncertainty leads to a substantial amount of asymptotic disagreement.
Number of Pages in PDF File: 44
Keywords: asymptotic disagreement, Bayesian learning, merging of opinions
JEL Classification: C11, C72, D83
Date posted: March 27, 2008 ; Last revised: August 28, 2008
© 2015 Social Science Electronic Publishing, Inc. All Rights Reserved.
This page was processed by apollo8 in 0.312 seconds