In this article, we study a social learning model in which agents iteratively update their beliefs about the true state of the world using private signals and the beliefs of… Click to show full abstract
In this article, we study a social learning model in which agents iteratively update their beliefs about the true state of the world using private signals and the beliefs of other agents in a non-Bayesian manner. Some agents are stubborn, meaning they attempt to convince others of an erroneous true state (modeling fake news). We show that while agents learn the true state on short timescales, they “forget” it and believe the erroneous state to be true on longer timescales. Using these results, we devise strategies for seeding stubborn agents so as to disrupt learning, which outperforms intuitive heuristics and gives novel insights regarding vulnerabilities in social learning.
               
Click one of the above tabs to view related content.