Learning Scene-Independent Group Descriptors for Crowd Understanding

Jing Shao, Chen Change Loy, Xiaogang Wang

Research output: Contribution to journalArticlepeer-review

63 Citations (Scopus)

Abstract

Groups are the primary entities that make up a crowd. Understanding group-level dynamics and properties is thus scientifically important and practically useful in a wide range of applications, especially for crowd understanding. In this paper, we show that fundamental group-level properties, such as intra-group stability and inter-group conflict, can be systematically quantified by visual descriptors. This is made possible through learning a novel collective transition prior, which leads to a robust approach for group segregation in public spaces. From the former, we further devise a rich set of group-property visual descriptors. These descriptors are scene-independent and can be effectively applied to public scenes with a variety of crowd densities and distributions. Extensive experiments on hundreds of public scene video clips demonstrate that such property descriptors are complementary to each other, scene-independent, and they convey critical information on physical states of a crowd. The proposed group-level descriptors show promising results and potentials in multiple applications, including crowd dynamic monitoring, crowd video classification, and crowd video retrieval.

Original languageEnglish
Article number7428859
Pages (from-to)1290-1303
Number of pages14
JournalIEEE Transactions on Circuits and Systems for Video Technology
Volume27
Issue number6
DOIs
Publication statusPublished - Jun 2017
Externally publishedYes

Bibliographical note

Publisher Copyright:
© 1991-2012 IEEE.

ASJC Scopus Subject Areas

  • Media Technology
  • Electrical and Electronic Engineering

Keywords

  • Crowded scene understanding
  • group-property analysis
  • video analysis

Cite this