6 December 2024 Detecting sequential video forgery using spatiotemporal attention mechanisms
Upasana Singh, Sandeep Rathor, Manoj Kumar
Author Affiliations +
Abstract

The advancement of video forgery technologies has made it increasingly easy to create highly realistic yet forged video content, raising significant concerns about their potential misuse. While existing forgery detection methods primarily focus on single-step forgeries, user-friendly video editing tools now enable multiple sequential forgeries. Detecting these sequential alterations is essential for identifying forged videos and understanding the forgery process. We propose a novel approach for detecting sequential video forgery. Unlike traditional binary classification tasks, our method predicts sequences of forgery steps. Our framework employs the Swin Transformer for spatiotemporal relation extraction, enhanced by a spatiotemporal relationships attention (STR) module and hierarchical attention for sequential relation modeling. Experiments show our method outperforms state-of-the-art techniques, achieving significant improvements in fixed-Acc and adaptive-Acc metrics. This research sets a new benchmark for sequential video forgery detection, providing an effective framework and valuable dataset for the community. Our results demonstrate the method’s capability in detecting complex, multi-step forgeries in video content.

© 2024 SPIE and IS&T
Upasana Singh, Sandeep Rathor, and Manoj Kumar "Detecting sequential video forgery using spatiotemporal attention mechanisms," Journal of Electronic Imaging 33(6), 063042 (6 December 2024). https://doi.org/10.1117/1.JEI.33.6.063042
Received: 6 August 2024; Accepted: 18 November 2024; Published: 6 December 2024
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Video

Counterfeit detection

Transformers

Video coding

Windows

Semantic video

Data modeling

Back to Top