Video prediction based on temporal aggregation and recurrent propagation for surveillance videos

P, Mohana Priya and K, UlagaPriya (2025) Video prediction based on temporal aggregation and recurrent propagation for surveillance videos. MethodsX, 14. p. 103402. ISSN 22150161

[thumbnail of main.pdf] Text
main.pdf

Download (2MB)

Abstract

Video prediction is essential for recreating absent frames in video sequences while maintaining temporal and spatial coherence. This procedure, known as video inpainting, seeks to reconstruct missing segments by utilizing data from available frames. Frame interpolation, a fundamental component of this methodology, detects and produces intermediary frames between input sequences. The suggested methodology presents a Bidirectional Video Prediction Network (BVPN) for precisely forecasting absent frames that occur before, after, or between specified input frames. The BVPN framework incorporates temporal aggregation and recurrent propagation to improve forecast accuracy. Temporal aggregation employs a series of reference frames to generate absent content by harnessing existing spatial and temporal data, hence assuring seamless coherence. Recurrent propagation enhances temporal consistency by integrating pertinent information from prior time steps to progressively improve predictions. The timing of frames is constantly controlled through intermediate activations in the BVPN, allowing for accurate synchronization and improved temporal alignment. A fusion module integrates intermediate interpretations to generate cohesive final outputs. Experimental assessments indicate that the suggested method surpasses current state-of-the-art techniques in video inpainting and prediction, attaining enhanced smoothness and precision. Surveillance video datasets demonstrate substantial enhancements in predictive accuracy, highlighting the strength and efficacy of the suggested strategy in practical application.•The proposed method integrates bidirectional video prediction, temporal aggregation, and recurrent propagation to effectively reconstruct missing intermediate video frames with enhanced accuracy.•Comparative analysis using the UCF-Crime dataset demonstrates higher PSNR and SSIM values for the proposed method, indicating improved frame quality and temporal consistency over existing techniques.•This research provides a robust framework for future advancements in video frame prediction, contributing to applications in anomaly detection, surveillance, and video restoration.

Item Type: Article
Subjects: Computer Science Engineering > Computer Network
Domains: Computer Science Engineering
Depositing User: Mr IR Admin
Date Deposited: 20 Aug 2025 10:02
Last Modified: 20 Aug 2025 10:02
URI: https://ir.vistas.ac.in/id/eprint/10120

Actions (login required)

View Item
View Item