The spread of disinformation on social media platforms is harmful to society. This harm may manifest as a gradual degradation of public discourse; but it can also take the form of sudden dramatic events such as the recent insurrection on Capitol Hill. The platforms themselves are in the best position to prevent the spread of disinformation, as they have the best access to relevant data and the expertise to use it. However, mitigating disinformation is costly, not only for implementing detection algorithms or employing manual effort, but also because limiting such highly viral content impacts user engagement and thus potential advertising revenue. Since the costs of harmful content are borne by other entities, the platform will therefore have no incentive to exercise the socially-optimal level of effort. This problem is similar to that of environmental regulation, in which the costs of adverse events are not directly borne by a firm, the mitigation effort of a firm is not observable, and the causal link between a harmful consequence and a specific failure is difficult to prove. For environmental regulation, one solution is to perform costly monitoring to ensure that the firm takes adequate precautions according to a specified rule. However, a fixed rule for classifying disinformation becomes less effective over time, as bad actors can learn to sequentially and strategically bypass it. Encoding our domain as a Markov decision process, we demonstrate that no penalty based on a static rule, no matter how large, can incentivize adequate effort. Penalties based on an adaptive rule can incentivize optimal effort, but counterintuitively, only if the regulator sufficiently overreacts to harmful events by requiring a greater-than-optimal level of effort. We prescribe the design of mechanisms that elicit platforms' costs of precautionary effort relating to the control of disinformation.
翻译:在社交媒体平台上散布虚假信息对社会有害。这种伤害可能表现为公共言论逐渐退化;但也可能表现为公共言论逐渐退化;但也可能表现为突发性戏剧性事件,如国会山最近发生的叛乱。平台本身最有能力防止虚假信息传播,因为它们最有机会获得相关数据,也最有能力使用这些数据。然而,减少错误信息的代价高昂,不仅用于执行检测算法或人工操作,而且因为限制这种高传播内容影响用户参与,从而影响潜在广告收入。由于有害内容的成本由其他实体承担,因此平台将没有动力来实施社会上最优化的努力。 这一问题与环境监管相似,因为环境监管的成本并非直接由企业承担,而减少错误事件的成本,企业的缓解努力是无法预见的,而有害后果和具体失败之间的因果关系也难以证明。对于环境监管而言,一种解决办法是进行成本高昂的监测,以确保企业根据具体规则采取足够的防范措施。然而,对错误信息进行分类的固定规则将比时间低,但要求做出更优化的防范努力,因为糟糕的行为者在战略上学习一个不精确的指令。