The Relationship of Digital Literacy, Exposure to AI-Generated Deepfake Videos, and the Ability to Identify Deepfakes in Generation X
Main Article Content
Abstract
The study explores the relationship between digital literacy, exposure to AI-generated deepfake videos, and the ability to identify deepfakes by Generation X in Indonesia who are currently between the ages of 43 and 58. It also analyzes the impact of deepfake identification capabilities on the cognitive, affective, and behavioral aspects of internet users. Through a survey involving 199 respondents taken from a total population of 42 million Generation X internet users in Indonesia, it applied a random sampling method. The sample size was determined by the Slovin formula with a confidence level of 90% and a margin of error of 7.1%. The descriptive analysis shows a moderate level of digital literacy and relatively low exposure to deepfakes. However, the ability to identify deepfakes was found to be low. The results of inferential statistical analysis show that digital literacy and exposure to deepfakes do not have a significant influence on the ability to identify deepfakes. Additionally, the ability to identify deepfakes does not significantly affect cognition, compassion, or behavior. While digital literacy is important, these findings reinforce the assumptions of Generation Theory and Media Dependency Theory. Additionally, it suggests that specific training on media manipulation technologies is needed to improve deepfake detection capabilities. This research implies that efforts to improve digital literacy should be expanded, including technical skills and critical thinking relevant to manipulative media such as deepfakes.
Article Details
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
The proposed policy for journals that offer open access
Authors who publish with this journal agree to the following terms:
- Copyright on any article is retained by the author(s).
- Author grant the journal, right of first publication with the work simultaneously licensed under a Creative Commons Attribution License that allows others to share the work with an acknowledgement of the work’s authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal’s published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work.
- The article and any associated published material is distributed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License
References
Ahmed, S. (2021). Fooled by the fakes: Cognitive differences in perceived claim accuracy and sharing intention of non-political deepfakes. Personality and Individual Differences, 182, 111074. https://doi.org/10.1016/j.paid.2021.111074
Ahmed, S. (2023). Navigating the maze: Deepfakes, cognitive ability, and social media news skepticism. New Media & Society, 25(5), 1108–1129. https://doi.org/10.1177/14614448211019198
Ahmed, S., Ng, S. W. T., & Bee, A. W. T. (2023). Understanding the role of fear of missing out and deficient self-regulation in sharing of deepfakes on social media: Evidence from eight countries. Frontiers in Psychology, 14. https://doi.org/10.3389/fpsyg.2023.1127507
Ameen, N., Hosany, S., & Taheri, B. (2023). Generation Z’s psychology and new‐age technologies: Implications for future research. Psychology & Marketing, 40(10), 2029–2040. https://doi.org/10.1002/mar.21868
Barari, S., Munger, K., & Lucas, C. (2024). Political Deepfakes are as Credible as Other Fake Media and (Sometimes) Real Media. The Journal of Politics, 732990. https://doi.org/10.1086/732990
Blancaflor, E., Anave, D. M., Cunanan, T. S., Frias, T., & Velarde, L. N. (2023). A Literature Review of the Legislation and Regulation of Deepfakes in the Philippines. Proceedings of the 2023 14th International Conference on E-Business, Management and Economics, 392–397. https://doi.org/10.1145/3616712.3616722
Burnham, S. L. F., & Arbeit, M. R. (2023). Social Media Literacy to Confront Far-Right Content: Saying “No” to Neutrality. Human Development, 67(3), 117–134. https://doi.org/10.1159/000531765
Caldwell, M., Andrews, J. T. A., Tanay, T., & Griffin, L. D. (2020). AI-enabled future crime. Crime Science, 9(1), 14. https://doi.org/10.1186/s40163-020-00123-8
Cetindamar, D., Abedin, B., & Shirahada, K. (2024). The Role of Employees in Digital Transformation: A Preliminary Study on How Employees’ Digital Literacy Impacts Use of Digital Technologies. IEEE Transactions on Engineering Management, 71, 7837–7848. https://doi.org/10.1109/TEM.2021.3087724
Chadwick, A., & Stanyer, J. (2022). Deception as a Bridging Concept in the Study of Disinformation, Misinformation, and Misperceptions: Toward a Holistic Framework. Communication Theory, 32(1), 1–24. https://doi.org/10.1093/ct/qtab019
Chen, H., Lin, Y., Li, B., & Tan, S. (2023). Learning Features of Intra-Consistency and Inter-Diversity: Keys Toward Generalizable Deepfake Detection. IEEE Transactions on Circuits and Systems for Video Technology, 33(3), 1468–1480. IEEE Transactions on Circuits and Systems for Video Technology. https://doi.org/10.1109/TCSVT.2022.3209336
Cinelli, M., Etta, G., Avalle, M., Quattrociocchi, A., Di Marco, N., Valensise, C., Galeazzi, A., & Quattrociocchi, W. (2022). Conspiracy theories and social media platforms. Current Opinion in Psychology, 47, 101407. https://doi.org/10.1016/j.copsyc.2022.101407
Diakopoulos, N., & Johnson, D. (2021). Anticipating and addressing the ethical implications of deepfakes in the context of elections. New Media & Society, 23(7), 2072–2098. https://doi.org/10.1177/1461444820925811
Dobber, T., Metoui, N., Trilling, D., Helberger, N., & De Vreese, C. (2021). Do (Microtargeted) Deepfakes Have Real Effects on Political Attitudes? The International Journal of Press/Politics, 26(1), 69–91. https://doi.org/10.1177/1940161220944364
Dourado, T. (2023). Who Posts Fake News? Authentic and Inauthentic Spreaders of Fabricated News on Facebook and Twitter. Journalism Practice, 17(10), 2103–2122. https://doi.org/10.1080/17512786.2023.2176352
Eiserbeck, A., Maier, M., Baum, J., & Abdel Rahman, R. (2023). Deepfake smiles matter less—The psychological and neural impact of presumed AI-generated faces. Scientific Reports, 13(1), 16111. https://doi.org/10.1038/s41598-023-42802-x
Fangming Dai, & Li, Z. (2024). Research on 2D Animation Simulation Based on Artificial Intelligence and Biomechanical Modeling. EAI Endorsed Transactions on Pervasive Health and Technology, 10. https://doi.org/10.4108/eetpht.10.5907
Federspiel, F., Mitchell, R., Asokan, A., Umana, C., & McCoy, D. (2023). Threats by artificial intelligence to human health and human existence. BMJ Global Health, 8(5), e010435. https://doi.org/10.1136/bmjgh-2022-010435
Fosco, C., Josephs, E., Andonian, A., & Oliva, A. (2022). Deepfake Caricatures: Human-guided Motion Magnification Improves Deepfake Detection by Humans and Machines. Journal of Vision, 22(14), 4079. https://doi.org/10.1167/jov.22.14.4079
Godulla, A., Hoffmann, C. P., & Seibert, D. M. A. (2021). Dealing with deepfakes—An interdisciplinary examination of the state of research and implications for communication studies. Studies in Communication and Media, 10(1), 73–96. Scopus. https://doi.org/10.5771/2192-4007-2021-1-72
Goh, D. H. (2024). “He looks very real”: Media, knowledge, and search‐based strategies for deepfake identification. Journal of the Association for Information Science and Technology, 75(6), 643–654. https://doi.org/10.1002/asi.24867
Guess, A. M., & Munger, K. (2023). Digital literacy and online political behavior. Political Science Research and Methods, 11(1), 110–128. https://doi.org/10.1017/psrm.2022.17
Hameleers, M., Van Der Meer, T. G. L. A., & Dobber, T. (2022). You Won’t Believe What They Just Said! The Effects of Political Deepfakes Embedded as Vox Populi on Social Media. Social Media + Society, 8(3), 205630512211163. https://doi.org/10.1177/20563051221116346
Hameleers, M., Van Der Meer, T. G. L. A., & Dobber, T. (2024). They Would Never Say Anything Like This! Reasons To Doubt Political Deepfakes. European Journal of Communication, 39(1), 56–70. https://doi.org/10.1177/02673231231184703
Hancock, J. T., & Bailenson, J. N. (2021). The Social Impact of Deepfakes. Cyberpsychology, Behavior, and Social Networking, 24(3), 149–152. https://doi.org/10.1089/cyber.2021.29208.jth
Harris, K. R. (2021). Video on demand: What deepfakes do and how they harm. Synthese, 199(5–6), 13373–13391. https://doi.org/10.1007/s11229-021-03379-y
Ienca, M. (2023). On Artificial Intelligence and Manipulation. Topoi, 42(3), 833–842. https://doi.org/10.1007/s11245-023-09940-3
Ismail, A., Elpeltagy, M., Zaki, M. S., & Eldahshan, K. (2022). An integrated spatiotemporal-based methodology for deepfake detection. Neural Computing and Applications, 34(24), 21777–21791. https://doi.org/10.1007/s00521-022-07633-3
Karpinska-Krakowiak, M. (Mags), & Eisend, M. (2024). Realistic Portrayals of Untrue Information: The Effects of Deepfaked Ads and Different Types of Disclosures. Journal of Advertising, 1–11. https://doi.org/10.1080/00913367.2024.2306415
Law, N., Woo, D., De La Torre, J., & Wong, G. (2018). A global framework of reference on digital literacy skills for indicator 4.4.2 (UIS/2018/ICT/IP/51). UNESCO Institute for Statistics.
Li, M., & Wan, Y. (2023). Norms or fun? The influence of ethical concerns and perceived enjoyment on the regulation of deepfake information. Internet Research, 33(5), 1750–1773. https://doi.org/10.1108/INTR-07-2022-0561
Lissitsa, S. (2024). Generations X, Y, Z: The effects of personal and positional inequalities on critical thinking digital skills. Online Information Review. https://doi.org/10.1108/OIR-09-2023-0453
Liu, Y., Chen, W., Liu, L., & Lew, M. S. (2019). SwapGAN: A Multistage Generative Approach for Person-to-Person Fashion Style Transfer. IEEE Transactions on Multimedia, 21(9), 2209–2222. https://doi.org/10.1109/TMM.2019.2897897
Long, T. Q., Hoang, T. C., & Simkins, B. (2023). Gender gap in digital literacy across generations: Evidence from Indonesia. Finance Research Letters, 58, 104588. https://doi.org/10.1016/j.frl.2023.104588
Lorenz-Spreen, P., Geers, M., Pachur, T., Hertwig, R., Lewandowsky, S., & Herzog, S. M. (2021). Boosting people’s ability to detect microtargeted advertising. Scientific Reports, 11(1), 15541. https://doi.org/10.1038/s41598-021-94796-z
Lorenz-Spreen, P., Lewandowsky, S., Sunstein, C. R., & Hertwig, R. (2020). How behavioural sciences can promote truth, autonomy and democratic discourse online. Nature Human Behaviour, 4(11), 1102–1109. https://doi.org/10.1038/s41562-020-0889-7
Lu, W., Liu, L., Zhang, B., Luo, J., Zhao, X., Zhou, Y., & Huang, J. (2023). Detection of Deepfake Videos Using Long-Distance Attention. IEEE Transactions on Neural Networks and Learning Systems, 1–14. IEEE Transactions on Neural Networks and Learning Systems. https://doi.org/10.1109/TNNLS.2022.3233063
MacLean, P., Cahillane, M., & Smy, V. (2024). Lifting the lid on manipulative website contents: A framework mapping contextual and informational feature combinations against associated social cognitive vulnerabilities. Social and Personality Psychology Compass, 18(2), e12947. https://doi.org/10.1111/spc3.12947
Malik, A., Kuribayashi, M., Abdullahi, S. M., & Khan, A. N. (2022). DeepFake Detection for Human Face Images and Videos: A Survey. IEEE Access, 10, 18757–18775. https://doi.org/10.1109/ACCESS.2022.3151186
Marron, M. B. (2015). New Generations Require Changes Beyond the Digital. Journalism & Mass Communication Educator, 70(2), 123–124. https://doi.org/10.1177/1077695815588912
McCosker, A. (2022). Making sense of deepfakes: Socializing AI and building data literacy on GitHub and YouTube. New Media and Society. Scopus. https://doi.org/10.1177/14614448221093943
Millière, R. (2022). Deep learning and synthetic media. Synthese, 200(3), 231. https://doi.org/10.1007/s11229-022-03739-2
Mo, S., Lu, P., & Liu, X. (2022). AI-Generated Face Image Identification with Different Color Space Channel Combinations. Sensors, 22(21), Article 21. https://doi.org/10.3390/s22218228
Mustak, M., Salminen, J., Mäntymäki, M., Rahman, A., & Dwivedi, Y. K. (2023). Deepfakes: Deceptions, mitigations, and opportunities. Journal of Business Research, 154, 113368. https://doi.org/10.1016/j.jbusres.2022.113368
Nas, E., & De Kleijn, R. (2024). Conspiracy thinking and social media use are associated with ability to detect deepfakes. Telematics and Informatics, 87, 102093. https://doi.org/10.1016/j.tele.2023.102093
Naskar, G., Mohiuddin, S., Malakar, S., Cuevas, E., & Sarkar, R. (2024). Deepfake detection using deep feature stacking and meta-learning. Heliyon, 10(4), e25933. https://doi.org/10.1016/j.heliyon.2024.e25933
Neethirajan, S. (2021). Is Seeing Still Believing? Leveraging Deepfake Technology for Livestock Farming. Frontiers in Veterinary Science, 8, 740253. https://doi.org/10.3389/fvets.2021.740253
Nieweglowska, M., Stellato, C., & Sloman, S. A. (2023). Deepfakes: Vehicles for Radicalization, Not Persuasion. Current Directions in Psychological Science, 32(3), 236–241. https://doi.org/10.1177/09637214231161321
Patel, Y., Tanwar, S., Bhattacharya, P., Gupta, R., Alsuwian, T., Davidson, I. E., & Mazibuko, T. F. (2023). An Improved Dense CNN Architecture for Deepfake Image Detection. IEEE Access, 11, 22081–22095. IEEE Access. https://doi.org/10.1109/ACCESS.2023.3251417
Prezja, F., Paloneva, J., Pölönen, I., Niinimäki, E., & Äyrämö, S. (2022). DeepFake knee osteoarthritis X-rays from generative adversarial neural networks deceive medical experts and offer augmentation potential to automatic classification. Scientific Reports, 12(1), 18573. https://doi.org/10.1038/s41598-022-23081-4
Qureshi, J., & Khan, S. (2024). Deciphering Deception—The Impact of AI Deepfakes on Human Cognition and Emotion. Journal of Advances in Artificial Intelligence, 2(1). https://doi.org/10.18178/JAAI.2024.2.1.101-107
Schmitt, K. L., & Woolf, K. D. (2018). Cognitive development in digital contexts. Journal of Children and Media, 1–3. https://doi.org/10.1080/17482798.2018.1522116
Shahzad, H. F., Rustam, F., Flores, E. S., Luís Vidal Mazón, J., De La Torre Diez, I., & Ashraf, I. (2022). A Review of Image Processing Techniques for Deepfakes. Sensors, 22(12), 4556. https://doi.org/10.3390/s22124556
Shin, S. Y., & Lee, J. (2022). The Effect of Deepfake Video on News Credibility and Corrective Influence of Cost-Based Knowledge about Deepfakes. Digital Journalism, 10(3), 412–432. https://doi.org/10.1080/21670811.2022.2026797
Sivathanu, B., Pillai, R., & Metri, B. (2023). Customers’ online shopping intention by watching AI-based deepfake advertisements. International Journal of Retail & Distribution Management, 51(1), 124–145. https://doi.org/10.1108/IJRDM-12-2021-0583
Tahir, R., Batool, B., Jamshed, H., Jameel, M., Anwar, M., Ahmed, F., Zaffar, M. A., & Zaffar, M. F. (2021). Seeing is Believing: Exploring Perceptual Differences in DeepFake Videos. Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, 1–16. https://doi.org/10.1145/3411764.3445699
Tinmaz, H., Lee, Y.-T., Fanea-Ivanovici, M., & Baber, H. (2022). A systematic review on digital literacy. Smart Learning Environments, 9(1), 21. https://doi.org/10.1186/s40561-022-00204-y
Tolosana, R., Vera-Rodriguez, R., Fierrez, J., Morales, A., & Ortega-Garcia, J. (2020). Deepfakes and beyond: A Survey of face manipulation and fake detection. Information Fusion, 64, 131–148. https://doi.org/10.1016/j.inffus.2020.06.014
Trinh, L., Tsang, M., Rambhatla, S., & Liu, Y. (2021). Interpretable and Trustworthy Deepfake Detection via Dynamic Prototypes. 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), 1972–1982. https://doi.org/10.1109/WACV48630.2021.00202
Twomey, J., Ching, D., Aylett, M. P., Quayle, M., Linehan, C., & Murphy, G. (2023). Do deepfake videos undermine our epistemic trust? A thematic analysis of tweets that discuss deepfakes in the Russian invasion of Ukraine. PLOS ONE, 18(10), e0291668. https://doi.org/10.1371/journal.pone.0291668
Vaccari, C., & Chadwick, A. (2020). Deepfakes and Disinformation: Exploring the Impact of Synthetic Political Video on Deception, Uncertainty, and Trust in News. Social Media + Society, 6(1), 205630512090340. https://doi.org/10.1177/2056305120903408
Van Der Sloot, B., & Wagensveld, Y. (2022). Deepfakes: Regulatory challenges for the synthetic society. Computer Law & Security Review, 46, 105716. https://doi.org/10.1016/j.clsr.2022.105716
Vasist, P. N., & Krishnan, S. (2023). Engaging with deepfakes: A meta-synthesis from the perspective of social shaping of technology theory. Internet Research, 33(5), 1670–1726. https://doi.org/10.1108/INTR-06-2022-0465
Waqas, N., Safie, S. I., Kadir, K. A., Khan, S., & Kaka Khel, M. H. (2022). DEEPFAKE Image Synthesis for Data Augmentation. IEEE Access, 10, 80847–80857. IEEE Access. https://doi.org/10.1109/ACCESS.2022.3193668
Xia, C., & Johnson, N. F. (2024). Nonlinear spreading behavior across multi-platform social media universe. Chaos: An Interdisciplinary Journal of Nonlinear Science, 34(4), 043149. https://doi.org/10.1063/5.0199655
Xiao, S., Zhang, Z., Yang, J., Wen, J., & Li, Y. (2023). Forgery Detection by Weighted Complementarity between Significant Invariance and Detail Enhancement. ACM Transactions on Multimedia Computing, Communications, and Applications. https://doi.org/10.1145/3605893
Yang, Z., Liang, J., Xu, Y., Zhang, X.-Y., & He, R. (2023). Masked Relation Learning for DeepFake Detection. IEEE Transactions on Information Forensics and Security, 18, 1696–1708. IEEE Transactions on Information Forensics and Security. https://doi.org/10.1109/TIFS.2023.3249566
Zhao, C., Wang, C., Hu, G., Chen, H., Liu, C., & Tang, J. (2023). ISTVT: Interpretable Spatial-Temporal Video Transformer for Deepfake Detection. IEEE Transactions on Information Forensics and Security, 18, 1335–1348. IEEE Transactions on Information Forensics and Security. https://doi.org/10.1109/TIFS.2023.3239223