Unsupervised Learning of Event Classes from Video

Authors

  • Muralikrishna Sridhar University of Leeds
  • Anthony Cohn University of Leeds
  • David Hogg University of Leeds

DOI:

https://doi.org/10.1609/aaai.v24i1.7726

Keywords:

unsupervised learning of event classes, video activity understanding, scene interpretation, video activity analysis, learning from observation

Abstract

We present a method for unsupervised learning of event classes from videos in which multiple actions might occur simultaneously. It is assumed that all such activities are produced from an underlying set of event class generators. The learning task is then to recover this generative process from visual data. A set of event classes is derived from the most likely decomposition of the tracks into a set of labelled events involving subsets of interacting tracks. Interactions between subsets of tracks are modelled as a relational graph structure that captures qualitative spatio-temporal relationships between these tracks. The posterior probability of candidate solutions favours decompositions in which events of the same class have a similar relational structure, together with other measures of well-formedness. A Markov Chain Monte Carlo (MCMC) procedure is used to efficiently search for the MAP solution. This search moves between possible decompositions of the tracks into sets of unlabelled events and at each move adds a close to optimal labelling (for this decomposition) using spectral clustering. Experiments on real data show that the discovered event classes are often semantically meaningful and correspond well with groundtruth event classes assigned by hand.

Downloads

Published

2010-07-05

How to Cite

Sridhar, M., Cohn, A., & Hogg, D. (2010). Unsupervised Learning of Event Classes from Video. Proceedings of the AAAI Conference on Artificial Intelligence, 24(1), 1631-1638. https://doi.org/10.1609/aaai.v24i1.7726