Fusion: Practice and Applications

Journal DOI

https://doi.org/10.54216/FPA

Submit Your Paper

2692-4048ISSN (Online) 2770-0070ISSN (Print)

Volume 20 , Issue 2 , PP: 77-92, 2025 | Cite this article as | XML | Html | PDF | Full Length Article

A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions

M. E. ElAlmi 1 * , M. M. Lotfy 2 * , M. M. Ghoniem 3 *

  • 1 Prof. of Computer and Information System, Faculty of Specific Education, Mansoura University, Egypt - (moh_elalmi@mans.edu.eg)
  • 2 Demonstrator of computer teacher preparation Department, Faculty of Specific Education, Mansoura University, Egypt - (mariamlotfy@mans.edu.eg)
  • 3 Lecturer of computer teacher preparation Department, Faculty of Specific Education, Mansoura University, Egypt - (m_ghonem@mans.edu.eg)
  • Doi: https://doi.org/10.54216/FPA.200207

    Received: January 05, 2025 Revised: March 09, 2025 Accepted: May 25, 2025
    Abstract

    Surveillance cameras play a pivotal role in educational institutions. They monitor the educational process, detect violations, and protect students from potential injuries or dangers. Continuous recording generates a massive amount of video data. Human observers spend significant time and effort reviewing the footage. Reviewing aims to detect and quickly address abnormal events. Abnormal events are rare in educational environments. Observers may become bored during continuous monitoring. This may cause fatigue and loss of attention. To overcome these challenges, this paper proposes an intelligent system that combines summarization and abnormal event detection in surveillance video. It is divided into two stages: The first stage starts with the extraction of static, feature-based key frames that highlight the video's most significant content. In the second stage, Convolutional Autoencoder (CAE) network used to detect abnormal events from the key frames generated by the summary stage. The proposed system produces two separate videos: a general summary and a dedicated abnormal events video sent to the relevant individuals. The proposed system was tested on some benchmark datasets. The experimental results demonstrated that the proposed system was effective in reducing browsing time and effort, as well as in detecting abnormal events within an educational context.

    Keywords :

    Static summarization , Surveillance video , Convolutional Autoencoder (CAE) , Abnormal Events

    References

    [1]       M. M. Elzain, "Surveillance Cameras Technologies at Imam Abdul Rahman Bin Faisal University," Arab Journal for Scientific Publishing (AJSP), 2023.

    [2]       M. Malekar, "Detecting criminal activities of surveillance videos using deep learning," Int. J. Sci. Res. Comput. Sci., vol. 7, no. 1, 2021.

    [3]       A. Kushwaha, M. Khare, R. M. Bommisetty, and A. Khare, "Human activity recognition based on video summarization and deep convolutional neural network," The Comput. J., vol. 67, no. 8, pp. 2601-2609, 2024.

    [4]       M. Yarrarapu, N. Leelavathy, and D. Haritha, "Efficient video summarization through MobileNetSSD: a robust deep learning-based framework for efficient video summarization focused on objects of interest," Multimedia Tools Appl., pp. 1-26, 2024.

    [5]       M. Javaid, M. Maqsood, F. Aadil, J. Safdar, and Y. Kim, "An efficient method for underwater video summarization and object detection using YoLoV3," Intell. Autom. Soft Compute, vol. 35, no. 2, 2023.

    [6]       B. Köprü and E. Erzin, "Use of affective visual information for summarization of human-centric videos," IEEE Trans. Affective Comput., vol. 14, no. 4, pp. 3135-3148, 2022.

    [7]       S. Paulraj and S. Vairavasundaram, "Transformer-enabled weakly supervised abnormal event detection in intelligent video surveillance systems," Eng. Appl. Artif. Intell., vol. 139, p. 109496, 2025.

    [8]       A. Mumtaz, A. B. Sargano, and Z. Habib, "AnomalyNet: a spatiotemporal motion-aware CNN approach for detecting anomalies in real-world autonomous surveillance," The Visual Comput., pp. 1-22, 2024.

    [9]       P. Y. Ingle and Y. G. Kim, "Real-time abnormal object detection for video surveillance in smart cities," Sensors, vol. 22, no. 10, p. 3862, 2022.

    [10]    S. Saponara, A. Elhanashi, and A. Gagliardi, "Real-time video fire/smoke detection based on CNN in antifire surveillance systems," J. Real-Time Image Process., vol. 18, pp. 889-900, 2021.

    [11]    J. Jung, S. Park, H. Kim, C. Lee, and C. Hong, "Artificial intelligence-driven video indexing for rapid surveillance footage summarization and review," in Proc. Thirty-Third Int. Joint Conf. Artif. Intell., pp. 8687-8690, Aug. 2024.

    [12]    A. Sabha and A. Selwal, "CoSumNet: A video summarization-based framework for COVID-19 monitoring in crowded scenes," Artif. Intell. Med., vol. 139, p. 102544, 2023.

    [13]    K. Muchtar, M. R. Munggaran, A. Mahendra, K. Anwar, and C. Y. Lin, "A unified video summarization for video anomalies through deep learning," in Proc. 2022 IEEE Int. Conf. Multimedia Expo Workshops (ICMEW), pp. 1-4, 2022.

    [14]    J. Chaki, N. Dey, J. Chaki, and N. Dey, "Histogram-based image color features," in Image Color Feature Extraction Techniques: Fundamentals and Applications, pp. 29-41, 2021.

    [15]    A. R. Zubair and O. A. Alo, "Grey level co-occurrence matrix (GLCM) based second order statistics for image texture analysis," arXiv preprint arXiv: 2403.04038, 2024.

    [16]    R. Mahum, A. Irtaza, M. Nawaz, T. Nazir, M. Masood, S. Shaikh, and E. A. Nasr, "A robust framework to generate surveillance video summaries using combination of Zernike moments and R-transform and deep neural network," Multimedia Tools Appl., vol. 82, no. 9, pp. 13811-13835, 2023.

    [17]    Z. Sedaghatjoo, H. Hosseinzadeh, and B. S. Bigham, "Local Binary Pattern (LBP) optimization for feature extraction," arXiv preprint arXiv: 2407.18665, 2024.

    [18]    F. Tessari, K. Yao, and N. Hogan, "Surpassing cosine similarity for multidimensional comparisons: Dimension insensitive Euclidean metric (DIEM)," arXiv preprint arXiv: 2407.08623, 2024.

    [19]    H. Wang, Z. Han, X. Xiong, X. Song, and C. Shen, "Enhancing yarn quality wavelength spectrogram analysis: A semi-supervised anomaly detection approach with convolutional autoencoder," Machines, vol. 12, no. 5, p. 309, 2024.

    [20]    E. Şengönül, R. Samet, Q. Abu Al-Haija, A. Alqahtani, B. Alturki, and A. A. Alsulami, "An analysis of artificial intelligence techniques in surveillance video anomaly detection: A comprehensive survey," Applied Sciences, vol. 13, no. 8, p. 4956, 2023.

    [21]    S. E. F. De Avila, A. P. B. Lopes, A. Luz Jr, and A. A. Araújo, "VSUMM: A mechanism designed to produce static video summaries and a novel evaluation method," Pattern Recognit. Lett, vol. 32, no. 1, pp. 56-68, 2011.

    [22]    R. T. Ionescu, S. Smeureanu, B. Alexe, and M. Popescu, "Unmasking the abnormal events in video," in Proc. IEEE Int. Conf. Comput. Vis., pp. 2895-2903, 2017.

    [23]    R. Hannane, A. Elboushaki, and K. Afdel, "Efficient video summarization based on motion SIFT-distribution histogram," in Proc. 13th Int. Conf. Comput. Graph. Imaging Visualization (CGiV), pp. 312-317, Mar. 2016.

    [24]    Y. Bendraou, F. Essannouni, and A. Salam, "From local to global key-frame extraction based on important scenes using SVD of centrist features," Multimedia Tools Appl., vol. 78, no. 2, pp. 1441-1456, 2019.

    [25]    M. U. Sreeja and B. C. Kovoor, "A multi-stage deep adversarial network for video summarization with knowledge distillation," J. Ambient Intell. Humanized Comput., vol. 14, no. 8, pp. 9823-9838, 2022.

    [26]    M. Hasan, J. Choi, J. Neumann, A. K. Roy-Chowdhury, and L. S. Davis, "Learning temporal regularity in video sequences," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 733-742, 2016.

    [27]    W. Luo, W. Liu, and S. Gao, "Remembering history with convolutional LSTM for anomaly detection," in Proc. IEEE Int. Conf. Multimedia Expo (ICME), pp. 439-444, Jul. 2017.

    [28]    W. Liu, W. Luo, D. Lian, and S. Gao, "Future frame prediction for anomaly detection – A new baseline," in Proc. IEEE Conf. Comput. Vis. Pattern Recognit., pp. 6536-6545, 2018.

    [29]    E. Cruz-Esquivel and Z. J. Guzman-Zavaleta, "An examination on autoencoder designs for anomaly detection in video surveillance," IEEE Access, vol. 10, pp. 6208-6217, 2022.

    Cite This Article As :
    E., M.. , M., M.. , M., M.. A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions. Fusion: Practice and Applications, vol. , no. , 2025, pp. 77-92. DOI: https://doi.org/10.54216/FPA.200207
    E., M. M., M. M., M. (2025). A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions. Fusion: Practice and Applications, (), 77-92. DOI: https://doi.org/10.54216/FPA.200207
    E., M.. M., M.. M., M.. A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions. Fusion: Practice and Applications , no. (2025): 77-92. DOI: https://doi.org/10.54216/FPA.200207
    E., M. , M., M. , M., M. (2025) . A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions. Fusion: Practice and Applications , () , 77-92 . DOI: https://doi.org/10.54216/FPA.200207
    E. M. , M. M. , M. M. [2025]. A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions. Fusion: Practice and Applications. (): 77-92. DOI: https://doi.org/10.54216/FPA.200207
    E., M. M., M. M., M. "A Two-Stage System for Surveillance Video Summarization and Unsupervised Abnormal Event Detection in Educational Institutions," Fusion: Practice and Applications, vol. , no. , pp. 77-92, 2025. DOI: https://doi.org/10.54216/FPA.200207