The development of powerful deep learning technologies has brought about some negative effects to both society and individuals. One such issue is the emergence of fake media. To tackle the issue, we have organized the Trusted Media Challenge (TMC) to explore how Artificial Intelligence (AI) technologies could be leveraged to combat fake media. To enable further research, we are releasing the dataset that we had prepared from the TMC challenge, consisting of 4,380 fake and 2,563 real videos, with various video and/or audio manipulation methods employed to produce different types of fake media. All the videos in the TMC dataset are accompanied with audios and have a minimum resolution of 360p. The videos have various durations, background, illumination, and may contain perturbations that mimic transmission errors and compression. We have also carried out a user study to demonstrate the quality of the TMC dataset and to compare the performance of humans and AI models. The results showed that the TMC dataset can fool human participants in many cases, and the winning AI models of the Trusted Media Challenge outperformed humans. The TMC dataset is available for research purpose upon request via tmc-dataset@aisingapore.org.
翻译:开发强大的深层学习技术给社会和个人带来了一些负面影响。其中一个问题就是假媒体的出现。为了解决这个问题,我们组织了 " 信任媒体挑战 ",以探讨如何利用人工智能技术打击假媒体。为了进行进一步研究,我们正在发布我们从TMC挑战中准备的数据集,其中包括4 380个假视频和2 563个真实视频,并采用各种视频和/或音频操纵方法制作不同类型的假媒体。TMC数据集中的所有视频都配有音频,并至少有360p的分辨率。视频有不同的持续时间、背景、照明,并可能含有模拟传输错误和压缩的干扰。我们还进行了用户研究,以展示TMC数据集的质量,比较人类和AI模型的性能。结果显示,TMC数据集在许多情况下可以愚弄人类参与者,并且赢得了信任媒体挑战的AI模型,其人形形形形色。TMC数据集可以应请求通过电子邮件数据系统进行研究。