This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

proofread

Study shows humans do not easily detect deepfakes

Study shows humans do not easily detect deepfakes
Illustration of the two sets of videos presented to participants. Credit: Computers in Human Behavior (2023). DOI: 10.1016/j.chb.2023.107917

A study of people's ability to detect "deepfakes" has shown humans perform fairly poorly, even when given hints on how to identify video-based deceit.

Dr. Klaire Somoray and Dr. Dan J Miller from James Cook University have published a new study in Computers in Human Behavior. They state that high-quality videos, in which a person in an existing image or video is manipulated to have another person's likeness, can now be generated with ease.

"This has raised concerns about this technology being used for nefarious purposes such as creating political misinformation. For instance, in March 2022, a manipulated video of Ukrainian President Volodymyr Zelensky was circulated, in which Zelensky is depicted appealing for Ukrainian soldiers to surrender," said Dr. Somoray.

Dr. Somoray and Dr. Miller recruited more than 450 people and showed them 20 videos, 10 of which were real and 10 of which were deepfakes. Participants were then graded on their ability to judge which videos were real, and which were not.

Half of the volunteers were given training on how to spot a deepfake video.

"This includes paying attention to things such as lighting, whether the cheeks and forehead looked too smooth or wrinkly, whether the agedness of the skin was similar to the agedness of the hair and eyes and whether looked real," the researchers said.

On average, participants correctly identified approximately 12 out of 20 videos.

"The poorest performers correctly categorized 5 out of 20 videos and the best performers correctly categorized 19 out of 20. Teaching people detection strategies did not impact detection accuracy or detection confidence, nor did time spent per video, or the average number of page clicks on each video," said Dr. Somoray.

"The findings cast doubt on whether simply providing the public with strategies for detecting deepfakes can meaningfully improve detection.

"Also, worryingly, it appears that may be overly optimistic regarding their abilities to ascertain the authenticity of individual videos," said Dr. Miller.

More information: Klaire Somoray et al, Providing detection strategies to improve human detection of deepfakes: An experimental study, Computers in Human Behavior (2023). DOI: 10.1016/j.chb.2023.107917

Journal information: Computers in Human Behavior

Citation: Study shows humans do not easily detect deepfakes (2023, September 12) retrieved 2 May 2024 from https://phys.org/news/2023-09-humans-easily-deepfakes.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Deepfake 'news anchors' in pro-China footage: research

14 shares

Feedback to editors