Learning an image-based motion context for multiple people tracking

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Laura Leal-Taixé
  • Michele Fenzi
  • Alina Kuznetsova
  • Bodo Rosenhahn
  • Silvio Savarese

Research Organisations

External Research Organisations

  • ETH Zurich
  • Stanford University
View graph of relations

Details

Original languageEnglish
Title of host publicationProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
PublisherIEEE Computer Society
Pages3542-3549
Number of pages8
ISBN (electronic)9781479951178, 9781479951178
Publication statusPublished - 24 Sept 2014
Event27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014 - Columbus, United States
Duration: 23 Jun 201428 Jun 2014

Publication series

NameProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
ISSN (Print)1063-6919

Abstract

We present a novel method for multiple people tracking that leverages a generalized model for capturing interactions among individuals. At the core of our model lies a learned dictionary of interaction feature strings which capture relationships between the motions of targets. These feature strings, created from low-level image features, lead to a much richer representation of the physical interactions between targets compared to hand-specified social force models that previous works have introduced for tracking. One disadvantage of using social forces is that all pedestrians must be detected in order for the forces to be applied, while our method is able to encode the effect of undetected targets, making the tracker more robust to partial occlusions. The interaction feature strings are used in a Random Forest framework to track targets according to the features surrounding them. Results on six publicly available sequences show that our method outperforms state-of-the-art approaches in multiple people tracking.

Keywords

    image-based motion context, linear programming, multiple people tracking, social force model

ASJC Scopus subject areas

Cite this

Learning an image-based motion context for multiple people tracking. / Leal-Taixé, Laura; Fenzi, Michele; Kuznetsova, Alina et al.
Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society, 2014. p. 3542-3549 6909848 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Leal-Taixé, L, Fenzi, M, Kuznetsova, A, Rosenhahn, B & Savarese, S 2014, Learning an image-based motion context for multiple people tracking. in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition., 6909848, Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, IEEE Computer Society, pp. 3542-3549, 27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014, Columbus, United States, 23 Jun 2014. https://doi.org/10.1109/CVPR.2014.453
Leal-Taixé, L., Fenzi, M., Kuznetsova, A., Rosenhahn, B., & Savarese, S. (2014). Learning an image-based motion context for multiple people tracking. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (pp. 3542-3549). Article 6909848 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition). IEEE Computer Society. https://doi.org/10.1109/CVPR.2014.453
Leal-Taixé L, Fenzi M, Kuznetsova A, Rosenhahn B, Savarese S. Learning an image-based motion context for multiple people tracking. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society. 2014. p. 3542-3549. 6909848. (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition). doi: 10.1109/CVPR.2014.453
Leal-Taixé, Laura ; Fenzi, Michele ; Kuznetsova, Alina et al. / Learning an image-based motion context for multiple people tracking. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition. IEEE Computer Society, 2014. pp. 3542-3549 (Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition).
Download
@inproceedings{b16008bfae8841cfbbc0a15b31799b34,
title = "Learning an image-based motion context for multiple people tracking",
abstract = "We present a novel method for multiple people tracking that leverages a generalized model for capturing interactions among individuals. At the core of our model lies a learned dictionary of interaction feature strings which capture relationships between the motions of targets. These feature strings, created from low-level image features, lead to a much richer representation of the physical interactions between targets compared to hand-specified social force models that previous works have introduced for tracking. One disadvantage of using social forces is that all pedestrians must be detected in order for the forces to be applied, while our method is able to encode the effect of undetected targets, making the tracker more robust to partial occlusions. The interaction feature strings are used in a Random Forest framework to track targets according to the features surrounding them. Results on six publicly available sequences show that our method outperforms state-of-the-art approaches in multiple people tracking.",
keywords = "image-based motion context, linear programming, multiple people tracking, social force model",
author = "Laura Leal-Taix{\'e} and Michele Fenzi and Alina Kuznetsova and Bodo Rosenhahn and Silvio Savarese",
year = "2014",
month = sep,
day = "24",
doi = "10.1109/CVPR.2014.453",
language = "English",
series = "Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition",
publisher = "IEEE Computer Society",
pages = "3542--3549",
booktitle = "Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition",
address = "United States",
note = "27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014 ; Conference date: 23-06-2014 Through 28-06-2014",

}

Download

TY - GEN

T1 - Learning an image-based motion context for multiple people tracking

AU - Leal-Taixé, Laura

AU - Fenzi, Michele

AU - Kuznetsova, Alina

AU - Rosenhahn, Bodo

AU - Savarese, Silvio

PY - 2014/9/24

Y1 - 2014/9/24

N2 - We present a novel method for multiple people tracking that leverages a generalized model for capturing interactions among individuals. At the core of our model lies a learned dictionary of interaction feature strings which capture relationships between the motions of targets. These feature strings, created from low-level image features, lead to a much richer representation of the physical interactions between targets compared to hand-specified social force models that previous works have introduced for tracking. One disadvantage of using social forces is that all pedestrians must be detected in order for the forces to be applied, while our method is able to encode the effect of undetected targets, making the tracker more robust to partial occlusions. The interaction feature strings are used in a Random Forest framework to track targets according to the features surrounding them. Results on six publicly available sequences show that our method outperforms state-of-the-art approaches in multiple people tracking.

AB - We present a novel method for multiple people tracking that leverages a generalized model for capturing interactions among individuals. At the core of our model lies a learned dictionary of interaction feature strings which capture relationships between the motions of targets. These feature strings, created from low-level image features, lead to a much richer representation of the physical interactions between targets compared to hand-specified social force models that previous works have introduced for tracking. One disadvantage of using social forces is that all pedestrians must be detected in order for the forces to be applied, while our method is able to encode the effect of undetected targets, making the tracker more robust to partial occlusions. The interaction feature strings are used in a Random Forest framework to track targets according to the features surrounding them. Results on six publicly available sequences show that our method outperforms state-of-the-art approaches in multiple people tracking.

KW - image-based motion context

KW - linear programming

KW - multiple people tracking

KW - social force model

UR - http://www.scopus.com/inward/record.url?scp=84911365771&partnerID=8YFLogxK

U2 - 10.1109/CVPR.2014.453

DO - 10.1109/CVPR.2014.453

M3 - Conference contribution

AN - SCOPUS:84911365771

T3 - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition

SP - 3542

EP - 3549

BT - Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition

PB - IEEE Computer Society

T2 - 27th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014

Y2 - 23 June 2014 through 28 June 2014

ER -

By the same author(s)