The spread of disinformation on social platforms is harmful to society. This harm may manifest as a gradual degradation of public discourse; but it can also take the form of sudden dramatic events such as the 2021 insurrection on Capitol Hill. The platforms themselves are in the best position to prevent the spread of disinformation, as they have the best access to relevant data and the expertise to use it. However, mitigating disinformation is costly, not only for implementing detection algorithms or employing manual effort, but also because limiting such highly viral content impacts user engagement and potential advertising revenue. Since the costs of harmful content are borne by other entities, the platform will therefore have no incentive to exercise the socially-optimal level of effort. This problem is similar to that of environmental regulation, in which the costs of adverse events are not directly borne by a firm, the mitigation effort of a firm is not observable, and the causal link between a harmful consequence and a specific failure is difficult to prove. For environmental regulation, one solution is to perform costly monitoring to ensure that the firm takes adequate precautions according to a specified rule. However, a fixed rule for classifying disinformation becomes less effective over time, as bad actors can learn to sequentially and strategically bypass it. Encoding our domain as a Markov decision process, we demonstrate that no penalty based on a static rule, no matter how large, can incentivize optimal effort. Penalties based on an adaptive rule can incentivize optimal effort, but counter-intuitively, only if the regulator sufficiently overreacts to harmful events by requiring a greater-than-optimal level of effort. We offer novel insights for the effective regulation of social platforms, highlight inherent challenges, and discuss promising avenues for future work.
翻译:在社会平台上散布不实信息对社会有害。这种伤害可能表现为公共言论逐渐退化;但也可能表现为2021年国会山叛乱等突发性突发事件。平台本身最有能力防止不实信息传播,因为它们最有机会获得相关数据,也最有能力使用这些数据。然而,减少不实信息代价高昂,不仅用于执行检测算法或采用人工操作,而且用于限制这种高端内容影响用户参与和潜在的广告收入。由于有害内容的成本由其他实体承担,因此平台将没有动力在社会上追求最佳水平的努力。 这一问题类似于环境监管,在环境监管方面,不利事件的代价不是直接由企业承担,公司缓解努力是无法预见的,而且有害后果与具体失败之间的因果关系也难以证明。对于环境监管而言,一种解决办法是进行代价高昂的监控,以确保公司根据具体规则采取足够的防范措施。然而,对错误信息进行分类的固定规则将变得不那么在时间上有效,而对于内在努力水平。 这一问题与环境监管类似,因为环境监管的成本不是直接承担,我们从战略上和上看,我们无法绕过一个最坏的轨道上决定。