We propose a detection-free system for segmenting multiple interacting and deforming people in a video. People detectors often fail under close agent interaction, limiting the performance of detection based tracking methods. Motion information often fails to separate similarly moving agents or to group distinctly moving articulated body parts. We formulate video segmentation as graph partitioning in the trajectory domain. We classify trajectories as foreground or background based on trajectory saliencies, and use foreground trajectories as graph nodes. We incorporate object connectedness constraints into our trajectory weight matrix based on topology of foreground: we set repulsive weights between trajectories that belong to different connected components in any frame of their time intersection. Attractive weights are set between similarly moving trajectories. Information from foreground topology complements motion information and our spatiotemporal segments can be interpreted as connected moving entities rather than just trajectory groups of similar motion. All our cues are computed on trajectories and naturally encode large temporal context, which is crucial for resolving local in time ambiguities. We present results of our approach on challenging datasets outperforming by far the state of the art.
Description
IEEE Xplore Full-Text HTML : Detection free tracking: Exploiting motion and topology for segmenting and tracking under entanglement
%0 Conference Paper
%1 5995366
%A Fragkiadaki, K.
%A Shi, Jianbo
%B Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on
%D 2011
%K motion_analysis motion_segmentation tracking
%P 2073-2080
%R 10.1109/CVPR.2011.5995366
%T Detection free tracking: Exploiting motion and topology for segmenting and tracking under entanglement
%U http://ieeexplore.ieee.org/xpls/icp.jsp?arnumber=5995366
%X We propose a detection-free system for segmenting multiple interacting and deforming people in a video. People detectors often fail under close agent interaction, limiting the performance of detection based tracking methods. Motion information often fails to separate similarly moving agents or to group distinctly moving articulated body parts. We formulate video segmentation as graph partitioning in the trajectory domain. We classify trajectories as foreground or background based on trajectory saliencies, and use foreground trajectories as graph nodes. We incorporate object connectedness constraints into our trajectory weight matrix based on topology of foreground: we set repulsive weights between trajectories that belong to different connected components in any frame of their time intersection. Attractive weights are set between similarly moving trajectories. Information from foreground topology complements motion information and our spatiotemporal segments can be interpreted as connected moving entities rather than just trajectory groups of similar motion. All our cues are computed on trajectories and naturally encode large temporal context, which is crucial for resolving local in time ambiguities. We present results of our approach on challenging datasets outperforming by far the state of the art.
@inproceedings{5995366,
abstract = {We propose a detection-free system for segmenting multiple interacting and deforming people in a video. People detectors often fail under close agent interaction, limiting the performance of detection based tracking methods. Motion information often fails to separate similarly moving agents or to group distinctly moving articulated body parts. We formulate video segmentation as graph partitioning in the trajectory domain. We classify trajectories as foreground or background based on trajectory saliencies, and use foreground trajectories as graph nodes. We incorporate object connectedness constraints into our trajectory weight matrix based on topology of foreground: we set repulsive weights between trajectories that belong to different connected components in any frame of their time intersection. Attractive weights are set between similarly moving trajectories. Information from foreground topology complements motion information and our spatiotemporal segments can be interpreted as connected moving entities rather than just trajectory groups of similar motion. All our cues are computed on trajectories and naturally encode large temporal context, which is crucial for resolving local in time ambiguities. We present results of our approach on challenging datasets outperforming by far the state of the art.},
added-at = {2014-07-29T14:58:49.000+0200},
author = {Fragkiadaki, K. and Shi, Jianbo},
biburl = {https://www.bibsonomy.org/bibtex/2c371255480df0d88601244c66a58daee/alex_ruff},
booktitle = {Computer Vision and Pattern Recognition (CVPR), 2011 IEEE Conference on},
description = {IEEE Xplore Full-Text HTML : Detection free tracking: Exploiting motion and topology for segmenting and tracking under entanglement},
doi = {10.1109/CVPR.2011.5995366},
interhash = {7098595fb6ef76ffd34a80703cb1d582},
intrahash = {c371255480df0d88601244c66a58daee},
issn = {1063-6919},
keywords = {motion_analysis motion_segmentation tracking},
month = {June},
pages = {2073-2080},
timestamp = {2014-07-29T14:58:49.000+0200},
title = {Detection free tracking: Exploiting motion and topology for segmenting and tracking under entanglement},
url = {http://ieeexplore.ieee.org/xpls/icp.jsp?arnumber=5995366},
year = 2011
}