Where Do Deep Fakes Look? Synthetic Face Detection via Gaze Tracking.

作者: Umur A. Ciftci , Ilke Demir

DOI: 10.1145/3448017.3457387

关键词:

摘要: Following the recent initiatives for democratization of AI, deep fake generators have become increasingly popular and accessible, causing dystopian scenarios towards social erosion trust. A particular domain, such as biological signals, attracted attention detection methods that are capable exploiting authenticity signatures in real videos not yet faked by generative approaches. In this paper, we first propose several prominent eye gaze features fakes exhibit differently. Second, compile those into analyze compare videos, formulating geometric, visual, metric, temporal, spectral variations. Third, generalize formulation to problem a neural network, classify any video wild or real. We evaluate our approach on datasets, achieving 89.79\% accuracy FaceForensics++, 80.0\% Deep Fakes (in wild), 88.35\% CelebDF datasets. conduct ablation studies involving different features, architectures, sequence durations, post-processing artifacts. Our analysis concludes with 6.29\% improved over complex network architectures without proposed signatures.

参考文章(58)
Erroll Wood, Tadas Baltruaitis, Xucong Zhang, Yusuke Sugano, Peter Robinson, Andreas Bulling, Rendering of Eyes for Eye-Shape Registration and Gaze Estimation 2015 IEEE International Conference on Computer Vision (ICCV). pp. 3756- 3764 ,(2015) , 10.1109/ICCV.2015.428
D.W. Hansen, J.P. Hansen, M. Nielsen, A.S. Johansen, M.B. Stegmann, Eye typing using Markov and active appearance models workshop on applications of computer vision. pp. 132- 136 ,(2002) , 10.1109/ACV.2002.1182170
V. Conotter, E. Bodnari, G. Boato, H. Farid, Physiologically-based detection of computer generated faces in video 2014 IEEE International Conference on Image Processing (ICIP). pp. 248- 252 ,(2014) , 10.1109/ICIP.2014.7025049
C. Connolly, T. Fleiss, A study of efficiency and accuracy in the transformation from RGB to CIELAB color space IEEE Transactions on Image Processing. ,vol. 6, pp. 1046- 1048 ,(1997) , 10.1109/83.597279
Moritz Kassner, William Patera, Andreas Bulling, Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction ubiquitous computing. pp. 1151- 1160 ,(2014) , 10.1145/2638728.2641695
Michael J. Reale, Peng Liu, Lijun Yin, Shaun Canavan, Art Critic: Multisignal Vision and Speech Interaction System in a Gaming Context IEEE Transactions on Systems, Man, and Cybernetics. ,vol. 43, pp. 1546- 1559 ,(2013) , 10.1109/TCYB.2013.2271606
, Generative Adversarial Nets neural information processing systems. ,vol. 27, pp. 2672- 2680 ,(2014) , 10.3156/JSOFT.29.5_177_2
Dan Witzner Hansen, Arthur E.C. Pece, Eye tracking in the wild Computer Vision and Image Understanding. ,vol. 98, pp. 155- 181 ,(2005) , 10.1016/J.CVIU.2004.07.013
D.W. Hansen, Qiang Ji, In the Eye of the Beholder: A Survey of Models for Eyes and Gaze IEEE Transactions on Pattern Analysis and Machine Intelligence. ,vol. 32, pp. 478- 500 ,(2010) , 10.1109/TPAMI.2009.30
Justus Thies, Michael Zollhofer, Marc Stamminger, Christian Theobalt, Matthias Niessner, Face2Face: Real-Time Face Capture and Reenactment of RGB Videos computer vision and pattern recognition. pp. 2387- 2395 ,(2016) , 10.1109/CVPR.2016.262