Appearance-based tracking of persons with an omnidirectional vision sensorShow others and affiliations
2003 (English)In: IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, IEEE, 2003, Vol. 7, article id 4624346Conference paper, Published paper (Refereed)
Abstract [en]
This paper addresses the problem of tracking a moving person with a single, omnidirectional camera. An appearance-based tracking system is described which uses a self-acquired appearance model and a Kalman filter to estimate the position of the person. Features corresponding to ``depth cues'' are first extracted from the panoramic images, then an artificial neural network is trained to estimate the distance of the person from the camera. The estimates are combined using a discrete Kalman filter to track the position of the person over time. The ground truth information required for training the neural network and the experimental analysis was obtained from another vision system, which uses multiple webcams and triangulation to calculate the true position of the person. Experimental results show that the tracking system is accurate and reliable, and that its performance can be further improved by learning multiple, person-specific appearance models
Place, publisher, year, edition, pages
IEEE, 2003. Vol. 7, article id 4624346
National Category
Computer Sciences
Research subject
Computer and Systems Science
Identifiers
URN: urn:nbn:se:oru:diva-4022DOI: 10.1109/CVPRW.2003.10072Scopus ID: 2-s2.0-84954442356ISBN: 0769519008 (print)OAI: oai:DiVA.org:oru-4022DiVA, id: diva2:138321
Conference
IEEE Workshop on Omnidirectional Vision, OMNIVIS 2003, Madison, Wisconsin, USA, June 21, 2003
2007-09-052007-09-052022-08-02Bibliographically approved