Deepfake detection by human crowds, machines, and machine-informed crowds
Abstract
The recent emergence of deepfake videos raises theoretical and practical questions. Are humans or the leading machine learning model more capable of detecting algorithmic visual manipulations of videos? How should content moderation systems be designed to detect and flag video-based misinformation? We present data showing that ordinary humans perform in the range of the leading machine learning model on a large set of minimal context videos. While we find that a system integrating human and model predictions is more accurate than either humans or the model alone, we show inaccurate model predictions often lead humans to incorrectly update their responses. Finally, we demonstrate that specialized face processing and the ability to consider context may specially equip humans for deepfake detection.
- Publication:
-
Proceedings of the National Academy of Science
- Pub Date:
- January 2022
- DOI:
- 10.1073/pnas.2110013119
- arXiv:
- arXiv:2105.06496
- Bibcode:
- 2022PNAS..11910013G
- Keywords:
-
- Computer Science - Computer Vision and Pattern Recognition;
- Computer Science - Artificial Intelligence
- E-Print:
- Proceedings of the National Academy of Sciences Jan 2022, 119 (1) e2110013119