Multi-feature-based crowd video modeling for visual event detection

Habib Ullah, Ihtesham Ul Islam*, Mohib Ullah, Muhammad Afaq, Sultan Daud Khan, Javed Iqbal

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

22 Scopus citations

Abstract

We propose a novel method for modeling crowd video dynamics by adopting a two-stream convolutional architecture which incorporates spatial and temporal networks. Our proposed method cope with the key challenge of capturing the complementary information on appearance from still frames and motion between frames. In our proposed method, a motion flow field is obtained from the video through dense optical flow. We demonstrate that the proposed method trained on multi-frame dense optical flow achieves significant improvement in performance in spite of limited training data. We train and evaluate our proposed method on a benchmark crowd video dataset. The experimental results of our method show that it outperforms five reference methods. We have chosen these reference methods since they are the most relevant to our work.

Original languageEnglish
Pages (from-to)589-597
Number of pages9
JournalMultimedia Systems
Volume27
Issue number4
DOIs
StatePublished - Aug 2021
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 2020, Springer-Verlag GmbH Germany, part of Springer Nature.

Keywords

  • CNN
  • Crowd analysis
  • Deep learning
  • Video modeling

ASJC Scopus subject areas

  • Software
  • Information Systems
  • Media Technology
  • Hardware and Architecture
  • Computer Networks and Communications

Fingerprint

Dive into the research topics of 'Multi-feature-based crowd video modeling for visual event detection'. Together they form a unique fingerprint.

Cite this