Assessing Deepfake Detection Models: A Comparative Study for Misinformation Detection and Prevention

dc.contributor.authorLiyana Gamage, Chathura
dc.contributor.departmentfi=Tietotekniikan laitos|en=Department of Computing|
dc.contributor.facultyfi=Teknillinen tiedekunta|en=Faculty of Technology|
dc.contributor.studysubjectfi=Tietotekniikka|en=Information and Communication Technology|
dc.date.accessioned2025-06-11T21:03:27Z
dc.date.available2025-06-11T21:03:27Z
dc.date.issued2025-06-02
dc.description.abstractDeepfake videos are proliferating faster than existing detection tools can keep pace, yet most prior studies benchmark detectors only on single datasets (with few exceptions that have generalized models), ignore bitrate degradation, and omit resource costs, leaving practitioners uncertain about real-world reliability. Addressing this gap, this study evaluates the reliability and deployability of selected deepfake detectors and formulates actionable countermeasures against synthetic media misinformation. Three state-of-the-art models, BA-TFD+, Convolutional Cross Efficient ViT, and CLRNet, were analysed on five publicly available datasets: DeeperForensics-1.0, Celeb-DF (v2), LAV-DF, DFD, and DFW under three H.264 compression settings. The extensive inference with measurement of precision, recall, AUC, F1-score, latency, and computational memory usage highlighted calibration gaps visualized through heat maps of AUC and F1 and AUC versus F1 scatter plots. Results show that at least one model achieved F1 approximately 0.80 on every dataset and compression levels and all detectors remain insensitive to bitrate reduction (i.e., higher compressed data), performance is fragmented as Cross ViT generalizes best but have a significant memory usage of 3-5 GB, BA-TFD+ offers near perfect accuracy on familiar dataset (i.e., LAV-DF) yet suffers severe threshold bias elsewhere, while CLRNet is lightweight and fast but not so well performed in terms of detection accuracy everywhere. This study concludes that no single detector rises as a winner in isolation. A combined or layered approach is recommended, such as edge-level pre-filters and cloud-side ensemble, continuous threshold calibration, and cryptographic watermarking techniques. Policy proposals include, but are not limited to, mandatory C2PA signatures, obligatory AI-labeling, and common standard compliance availability worldwide. Together, these technical and policy-wise measures provide a solid roadmap for sustaining trust in visual evidence as generative media evolves.
dc.format.extent124
dc.identifier.olddbid199023
dc.identifier.oldhandle10024/182061
dc.identifier.urihttps://www.utupub.fi/handle/11111/20067
dc.identifier.urnURN:NBN:fi-fe2025061166652
dc.language.isoeng
dc.rightsfi=Julkaisu on tekijänoikeussäännösten alainen. Teosta voi lukea ja tulostaa henkilökohtaista käyttöä varten. Käyttö kaupallisiin tarkoituksiin on kielletty.|en=This publication is copyrighted. You may download, display and print it for Your own personal use. Commercial use is prohibited.|
dc.rights.accessrightsavoin
dc.source.identifierhttps://www.utupub.fi/handle/10024/182061
dc.subjectdeepfake detection, domain generalization, misinformation mitigation, threshold calibration, cryptographic watermarking
dc.titleAssessing Deepfake Detection Models: A Comparative Study for Misinformation Detection and Prevention
dc.type.ontasotfi=Diplomityö|en=Master's thesis|

Tiedostot

Näytetään 1 - 1 / 1
Ladataan...
Name:
LiyanaGamage_Chathura_Thesis.pdf
Size:
942.59 KB
Format:
Adobe Portable Document Format