key: cord-0205360-mp7c0rpw authors: Macault, Emilien title: Stochastic Consensus and the Shadow of Doubt date: 2022-01-28 journal: nan DOI: nan sha: 19e8706edd14a79179be9ecec499cfc2150cfeee doc_id: 205360 cord_uid: mp7c0rpw We propose a stochastic model of opinion exchange in networks. A finite set of agents is organized in a fixed network structure. There is a binary state of the world and each agent receives a private signal on the state. We model beliefs as urns where red balls represent one possible value of the state and blue balls the other value. The model revolves purely around communication and beliefs dynamics. Communication happens in discrete time and, at each period, agents draw and display one ball from their urn with replacement. Then, they reinforce their urns by adding balls of the colors drawn by their neighbors. We show that for any network structure, this process converges almost-surely to a stable state. Futher, we show that if the communication network is connected, this stable state is such that all urns have the same proportion of balls. This result strengthens the main convergence properties of non-Bayesian learning models. Yet, contrary to those models, we show that this limit proportion is a full-support random variable. This implies that an arbitrarily small proportion of misinformed agents can substantially change the value of the limit consensus. We propose a set of conjectures on the distribution of this limit proportion based on simulations. In particular, we show evidence that the limit belief follows a beta distribution and that its average value is independent from the network structure. Despite public investments in media education and the development of counter-measures over the past few years, misinformation remains an ongoing issue with tangible consequences. The recent examples of the COVID-19 pandemics or the US presidential elections have highlighted how quickly inaccurate, deceptive or politically biased information spreads in a context of distrust towards experts and institutions. Under pressure to limit the spread of deceptive content, media and open web companies have put in place a set of policies to regulate news contents on their platforms. Such policies mostly include source highlighting, fact checking and advertisement campaigns, all of which have proved to have a limited e ciency. The failure of counter-disinformation policies may be explained by theoretical modeling shortcomings. Indeed, most policies are based on the assumption that agents behave rationally when it comes to information processing. By displaying the limited trustworthiness level of a spurious source, they assume agents will revise their beliefs over secure sources and naturally evacuate false news. In terms of economic modeling, this translates into the assumption that agents behave in a Bayesian manner. In a multi-agent context, the computational limitations of Bayesian models have incentivized the emergence of non-Bayesian models as an alternative. Most of these models are based on DeGroot (1974) and consist in setups where agents communicate by repeatedly averaging their opinions with their neighbors' until a steady state is reached. The strength of this framework is that beliefs quickly converge to a tractable limit. Yet, two major limitations are to be opposed to models based on DeGroot dynamics. First, they have been shown to have limited robustness, in the sense that the repeated averaging overweights initial beliefs while enforcing fast convergence of beliefs. Second, one may question the relevance of a setup where agents directly access and exchange their beliefs on some state. In most smallworld communication setups, it seems more realistic to assume that agents do not access the subjective probabilities they put on the possible values of a state, but rather decide to relay some information over another according to the relative probabilities they put on those events. In other terms, they do not exchange beliefs but draws based on beliefs. In this paper, we introduce and analyze a stochastic variant of DeGroot dynamics where agents behave in this respect. Introducing some degree of stochasticity strongly changes the perspective on misinformation: in DeGroot (1974) , slightly modifying the prior beliefs of some agents cannot change drastically the consensus outcome. Yet, most disinformation platforms display some "shadow of doubt" strategy: agents do not transmit false informations because they necessarily believe them to be true, but rather because there is some -even small -probability that they may not be false. In other words, they manage to disinform by inducing limited beliefs on their false informations. 1.1. Contribution. In this paper, we build an opinion formation model where agents communicate by drawing states according to their beliefs instead of directly communicating subjective probabilities. To do so, we model beliefs using reinforcing urns. Studying the evolution of beliefs comes down to characterizing the evolution of urns' compositions. Using stochastic approximation techniques, we show that in such models, under very general conditions, the dynamics of beliefs converge to a rest point. We show that at the steady-state, all agents share the exact same belief on the state of the world. As long as initial beliefs cover the whole state space, the consensus is drawn from a distribution with full support. This strongly contradicts the predictions of DeGroot's and similar models. We then try to characterize this limit distribution using simulations. 1.2. Literature. The emergence of consensus and its connection to learning heuristics is a long lasting question in theoretical economics. Two approaches are generally opposed to this problem: Bayesian and non-Bayesian models. Bayesian models on the emergence of consensus mostly started with Aumann (1976) and its seminal result that two agents with equal prior beliefs and common knowledge posteriors must have equal beliefs. Generalizations have been proposed by Geanakoplos and Polemarchakis (1982) and Parikh and Krasucki (1990) who showed respectively that two players repeatedly communicating must agree in the long run and that a nite umber of players communicating in pairs will eventually agree. This is partly due to actions being observable, although Bala and Goyal (1998) proves that when agents are embedded in a connected network and observe the outcome of their actions with some noise, players are able to learn the true payo distributions for actions that their neighbors take in nitely often, hence all actions converge to a consensual action. It is worth noting that although Bala and Goyal (1998) considers Bayesian agents, they limit their ability to compute beliefs by assuming that they do not make inferences on unobserved players and behave myopically. These results are closely related to social learning models and the observational learning literature as Banerjee (1992) , Smith and Sørensen (2000) or Rosenberg et al. (2009) . Mossel et al. (2020) generalizes those results to a large class of social learning models by introducing the concept of social learning equilibrium to study the asymptotic properties of learning processes and characterize conditions that agreement and herding behavior. In the same line, and Acemoglu et al. (2014) connect the emergence of social learning with Bayesian agents and the topology of the communication network. Non-Bayesian consensus models emerged through DeGroot (1974) , where the author introduces a model where agents living in a network repeatedly exchange their beliefs over some state of the world. At each stage, each agent replaces his belief by the average of his neighbors' beliefs. It is shown that if the communication network is connected, beliefs converge to a consensus which depends on initial beliefs and the network topology only. Variations on the updating rule have been proposed, for instance in Friedkin and Johnsen (1997) where authors allow some persistence on agents beliefs by including one's own belief in the averaging process. DeGroot's model gained popularity with Golub and Jackson (2010) which connects consensus and learning with the network's adjacency matrix using properties of Markov chains steady-states. They refer to DeGroot's belief averaging dynamics as naive learning. Examples of the use of DeGroot (1974) in economic modeling are too numerous to be listed. In the recent literature, Mandel and Venel (2020) considers a stochastic game where two misinformers try to in uence a population of agents applying naive learning. We refer the reader to and Golub and Sadler (2017) for surveys on both Bayesian and non-Bayesian learning in networks. In this paper, we question the limits of DeGroot's model by allowing agents to communicate via draws made according to their beliefs. In this regard, this work is in line with the existing literature on robustness of learning dynamics. Acemoglu et al. (2016) questions the predictions of Bayesian learning models by introducing uncertainty on the distribution of private signals agents receive. Criticism on DeGroot dynamics already featured in Golub and Jackson (2010) where authors proved that in the general case, beliefs do not converge for countably in nite player sets. In a recent work, Peretz et al. (2021) shows that in the presence of agents with xed beliefs over time, which they call bots, the common limit can converge to any value. Our approach to this problem is based on the seminal Polya urn model from Eggenberger and Pólya (1923) . In this paper, authors consider an urn with balls of several colors and study the convergence of reinforcement dynamics. It is a well known result that proportions in the urn converge to a beta distribution (see Klenke (2013) for instance). The strength of Polya's model is its intricate connection with exchangeability. Central papers in the foundation of Bayesian inference like De Finetti (1929) and Hewitt and Savage (1955) heavily rely on the concept of exchangeability. Polya urn plays a particular role in that Hill et al. (1987) proved that any exchangeable process of {0, 1}-valued random variables if either Bernoulli, deterministic or generated by draws from a Polya urn. Our model considers a system of interacting urns in the avor of Paganoni and Secchi (2004) which introduced such systems and rst proved convergence when the number of balls in all urns grow at the same speed. Similarly, Dai Pra et al. (2014) shows convergence of proportions in a system where urn reinforcement depends both on proportions in each urn and on the average proportions in the system. Crimaldi et al. (2016) gives further results on convergence and uctuations around the limit of such system. The model we consider is close yet di erent, as we consider a system where urns are reinforced at di erent speeds which correspond to their degree in the communication network. Usual probabilistic tools do not apply as proportions are not martingales and draws are not exchangeable. Instead, we rely on stochastic approximation as introduced by Robbins and Monro (1951) . Motivation for the use of stochastic approximation in the study of urn systems can be found in Laruelle et al. (2013) , where authors use this technique in the context of clinical trial modelling. 2. T M 2.1. Model. We consider a nite population = {1, . . . } of agents, embedded in a exogeneous and xed undirected graph = ( , ) with edge set . We denote by ( ) the neighborhood of any player and the degree of . We denote by = ( ) , ∈ the adjacency matrix of , with the convention that = 0 for all ∈ . Throughout, we will assume that is connected that is, for every pair of nodes , in there exists a path in connecting to . We consider a binary state space Θ = { , }. At the beginning of the game, a state is drawn at random from Θ and is unobserved. Players initially receive a noisy signal informing them about the state of the world. Formally, with˜being the realized value of the state, player receives a signal =˜with probability =˜with probability 1 − With˜representing the complementary value of˜in Θ. For a large network, by the law of large numbers, represents the average proportion of agents initially well-informed. 2.2. Beliefs and communication. We model beliefs using urns of in nite capacity with balls of two colors representing the possible values of the state : white balls represent the event =ã nd black balls represent the complementary event. At any time ≥ 1, the proportions of balls in agent 's urn then represent 's belief over those two events. Urns are initialized with a ball corresponding to the agent's signal. We de ne the communication process as a discrete time dynamics. At each stage, players draw one ball with replacement from their urns with uniform probability. Draws are assumed to be pairwise independent. Every agent observes the colors drawn by their neighbors. Then, beliefs are updated by reinforcing urns, adding one ball of the corresponding color for each draw from their neighbors. The process is repeated in nitely. At any given time, an agent's current belief on is given by the proportions of balls in his urn. Our objective is to study the evolution of the urn system and determine whether proportions converge, if a consensus is reached and if so, to characterize it given the network topology and the value of . 2.3. Example. Consider = 3 agents connected in line as displayed in Fig. 1 . Assume that at time = 0, agent 1 and 3 received a truthful signal and agent 2 got the wrong one. Then urns 1 and 3 will contain a white ball and urn 2 will contain one blackball. At time = 1, every player draws the only ball their urns contain and display it. Then, they all replace their draw and add a new ball of the corresponding color for every draw they observe. That is, at the end of the rst stage, urns 1 and 3 will contain each one black ball and one white ball, and urn 2 will contain one black ball and two white balls. At time = 2, the draw and reinforcement procedure is repeated: the end urns will draw one black or one white ball with equal probability and the middle urn will draw one white ball with probability 2/3 and a black ball with probability 1/3. Table 1 details the possible outcomes at time = 2. The left column is the vector of draws from urn 1, 2 and 3 respectively and the right column gives the compositions at the end of the time period in the same order. Left gures correspond to the number of white balls in the urn and right gures to the number of black balls. Vector of draws Probability of occurence Urns compositions (W,W,W) 1/6 (2,1)-(4,1)-(2,1) (W,B,W) 1/12 (1,2)-(4,1)-(1,2) (B,W,W) 1/6 (2,1)-(3,2)-(2,1) (B,B,W) 1/12 (1,2)-(3,2)-(1,2) (W,W,B) 1/6 (2,1)-(3,2)-(2,1) (W,B,B) 1/12 (1,2)-(3,2)-(1,2) (B,W,B) 1/6 (2,1)-(1,4)-(2,1) (B,B,B) 1/12 (1,2)-(1,4)-(1,2) 3. R To ease the characterization of the dynamics, we introduce some notations. First, denote by and respectively the numbers of white and black balls in player 's urn at time . De ne = + the total number of balls in player 's urn at time . One has that +1 = + and 0 = 1 hence 1 + . Let = / be the proportion of black balls in urn after step and = ( ) ∈ . Finally, let be the indicator variable of a black draw for agent at time , = ( 1 , . . . , ) and let F be the sigma-eld generated by the realizations of ( ), ≤ . We derive the following dynamics: Eq. (3.2) shows that, in expectation, the belief updating process obeys some local averaging property as in canonical naive learning models: the variation of proportion in any urn evolves according to the proportions in the neighboring urns. 3.2. Convergence of beliefs. We rst show that beliefs converge in the sense that color proportions in each urn converge to a stable point. The proof relies on stochastic approximation techniques, as usual probability methods do not apply in our case. Indeed, unless the graph is regular, neither local nor global proportions behave as martingales and it is easy to see that the process ( ) is not exchangeable as the rate at which an urns evolve depends both its degree and time. Nevertheless, we are able to frame the dynamics as an algorithm for which we can prove convergence. Theorem 1. For any graph , lim →∞ exists almost-surely for any ∈ . Proof. From Eq. (3.1) we derive the following recursive formula on : By adding and subtracting the conditional expectation of the number of black draws in neighboring urns to the numerator, we have: Observing that, conditional on F , the expected number of black draws in neighboring urns at time + 1 is equal to the sum of their proportions at time , we obtain: We now rescale the equation by a factor that is independent of : Where = min . Finally, we obtain the following system: In order to ensure convergence of the stochastic system Eq. (3.7), we rst make the following observations. Assumption 1 is central in any stochastic approximation algorithm à la Robbins and Monro (1951) with deterministic weights. These weights serve as the increments of time discretization. In that perspective, the rst point implies that the algorithm will cover the entire time interval. The second point involves, jointly with the next observation, the disappearing of noise in the limit. As is of the order of 1/ , Assumption 1 is immediate. Assumption 2. For every in , the sequence ( ) is a martingale di erence noise relative to F . Assumption 2, when combined with the second point in Assumption 1, ensures that the cumulative error due to the discretization noise is negligible almost-surely, as the noise variance will vanish asymptotically. Assumption 2 holds as, for any ∈ , the sequence ( ) is a sequence of bounded random variables with zero mean. Assumption 3. The maps are Lipschitz continuous and measurable with respect to F and uniformly continuous in for ≥ 1. Stochastic approximation ensures that a discrete-time stochastic process evolves along the trajectories of a continuous time ordinary di erential equation. In that respect, Assumption 3 ensures that the ODE is well de ned and has a unique solution. Finally, although the maps ( ) , in Eq. (3.7) depend on time, for any ∈ , the sequence of maps ( ) converge to a time-independent limit as time goes to in nity. Indeed, for any ∈ and any ∈ [0, 1] , let¯( ) = Assumption 4 holds immediately as, for any ∈ , →¯as → ∞. Based on Assumptions 1-4, we can apply Theorem 2.3 from Kushner and Yin (2003, chapter 5) . Theorem 2 (Kushner and Yin (2003, chapter 5) ). If Assumptions 1-4 hold and is bounded with probability one, then for almost all , the limits¯( ) of convergent subsequences of ( ) are trajectories of =¯( ) (3.10) in some bounded invariant set and ( ) converges to this invariant set. This result ensures that the system Eq. (3.7) evolves almost-surely along trajectories of Eq. (3.10) and converges to the set of asymptotically stable points of the ordinary di erential system. This rst result ensures that for any graph structure and any initial condition on the urns, proportions converge almost-surely to a stable point. In particular, convergence is independent of the initial signal structure and applies for any alternative initialization of the system. The next result details when a consensus emerges. Theorem 3. Suppose that the graph is connected. Then for any , ∈ , lim →∞ = lim →∞ almost-surely. Proof. From Theorem 4, we know that proportions converge along the trajectories of Eq. (3.10), that is: As 1 ≤ ≤ 1, stable points of Eq. (3.10) belong to the set of stable points of Where is the Laplacian matrix of the graph , i.e. = − with the diagonal matrix of degrees. Thus, Lyapounov stable solutions of Eq. (3.10) belong to the nullspace of − , as is symmetric, positive semi-de nite. As is connected, this nullspace is of dimension 1 and is characterized by the eigenvector (1, ..., 1) as the sum of each row in equals zero. As − is negative, the entire set is Lyapounov stable. Theorem 4. For any connected graph , if ∈ (0, 1), then the limit belief¯is a non-trivial distribution with full support on [0, 1]. Proof. The proof is based on the concept of attainability from Benaïm (1999) . De nition 5. A point ∈ R is attainable by if for every > 0 and every open neighborhood of , In other terms, a point is attainable if, from any vector of proportions, there is a strictly positive probability that becomes arbitrarily close to in nite time. Let¯denote the set of equilibrium points of Eq. (3.10) intersected with [0, 1] that is,¯= ∈ [0, 1] | = ∀ , ≤ . We establish the following lemma. To prove Lemma 6, simply observe that, from Eq. (3.1), +1 − is of the order of 1 . If ∈ (0, 1), P ∈ (0, 1) > 0 for every and ≥ 0. We showed that any point in¯is attainable. To complete the proof of Theorem 4, it remains to show that any attainable point in¯belongs to the support of ∞ . Observe that if = 0 or = 1, urns in the system display only one color hence beliefs will remain at their original value forever. The next section provides some empirical evidence on the limit distribution of beliefs. While our e orts in characterizing the limit distribution of the consensus as a function of and failed, large scale simulations provide some useful evidence. We simulated the learning dynamics on three network structures: stars, regular graphs with varying degree and complete networks. The values of the limit belief were simulated for di erent values of . Main elements of code used for the simulation feature in the appendix. 4.1. Evidence of a Beta Distribution. In the classical model from Eggenberger and Pólya (1923) , an urn is initialized at time = 0 with ≥ 1 white balls and ≥ 1 black balls. Then, at each discrete time step, a ball is drawn from the urn and replaced with ≥ 1 additional balls of the same color. It is widespread that the proportion of white balls converges in distribution to a beta distribution B( / , / ) (see Mahmoud (2008) ). For any two reals , > 0, the beta distribution B( , ) has a density function where ( , ) = Γ( )Γ( ) Γ( + ) and Γ is the Gamma function. Although we consider a system of interacting urns rather than a single urn, the beta distribution stands as a strong candidate for the limit distribution. Conjecture 7. The distribution of¯follows a beta distribution B( , ) for some , , > 0 which depend only on and . We collected the values of the limit belief for star, -regular and complete graphs of xed size = 100 and xed values of . As all the simulations were run independently, for any given graph structure and value of , the set of values of the limit beliefs is an i.i.d. sample. A beta distribution tting was computed by maximum likelihood estimation. Other distributions were tted in order to assess goodness-of-t using usual criteria. Fig. 2 compares the tted distributions assuming respectively a normal distribution and a beta distribution. The graph displays empirical and theoretical densities, quantile-quantile plots, cumulative distribution functions and probability plots. The beta distribution clearly appears as well tted to the sample. Additional plots feature in the appendix for di erent network structures and values of . In all the aforementioned cases, tness measures yielded similar results, where the beta distribution clearly appears as more adapted to describe the data. Estimation of the Parameters. Assuming the limit belief does follow a beta distribution, we are able to estimate its parameters using maximum likelihood estimation for various networks and values of . Our rst conjecture concerns the average of the limit distribution. All the simulations we ran conducted to a strong belief in that its value is . In other terms, the expected proportion of black balls in the limit is equal to the expected number of misinformed agents ex-ante. Conjecture 8. For any ∈ , →¯where¯∼ B( , ) with , > 0 such that + = . To support this conjecture, we simulated the communication dynamics on a 10-regular network for increasing values of alpha and tted a beta distribution to the empirical distribution. Similar simulations for other network structures yield identical results. These outputs strongly suggest that the belief updating dynamicsEq. (3.1) bear some form of asymptotical exchangeability. It remains to be proved yet it would strongly support Conjecture 7. Finally, we observe consistency in the sum of the estimatesˆandˆ, which support the following conjecture. Conjecture 9. Fix a network and let¯be the beta distribution of the limit belief given . Let and be its parameters. Then the mapping ↦ → + is constant. To support this conjecture, we explored results from simulations on the three aforementioned structures. We believe that discrepancies as observed in Table 2 are due to noise introduced by the random number generator as they mostly appear for extreme values of . Further work has been done in trying to characterize the limit distribution of the consensus analytically, yet no method has yield convincing results so far. Based on the simulations we ran, we have strong hints that suggest this limit distribution is a beta distribution. This con rms the intuition that, as a whole, the system acts as a global Polya urn. Further exploration of possible exchangeability properties might help in supporting this intuition and provide some tools for a closed-form characterization. To our knowledge, no paper has been able to achieve such formal results on interacting urn systems. A better understanding of the limit distribution would provide a better applicability of our results, in particular in designing a model of strategic disinformation with disinformants being That application was the initial motivation of the paper and remains its main objective Dynamics of information exchange in endogenous social networks Fragility of asymptotic agreement under bayesian learning Bayesian learning in social networks Opinion dynamics and learning in social networks Agreeing to disagree. The annals of statistics Learning from neighbours. The review of economic studies A simple model of herd behavior. The quarterly journal of economics Dynamics of stochastic approximation algorithms Fluctuation theorems for synchronization of interacting pólya's urns. Stochastic processes and their applications Synchronization via interacting reinforcement Funzione caratteristica di un fenomeno aleatorio Reaching a consensus Über die statistik verketteter vorgänge Social positions in in uence networks We can't disagree forever Naive learning in social networks and the wisdom of crowds Learning in social networks Symmetric measures on cartesian products Exchangeable urn processes Probability theory: a comprehensive course Stochastic Approximation and Recursive Algorithms and Applications. Stochastic Modelling and Applied Probability Randomized urn models revisited using stochastic approximation Pólya urn models Dynamic competition over social networks Social learning equilibria Interacting reinforced-urn systems Communication, consensus, and knowledge Robust naive learning in social networks A stochastic approximation method. The annals of mathematical statistics Informational externalities and emergence of consensus Pathological outcomes of observational learning