Previous [ 1] [ 2] [ 3] [ 4] [ 5] [ 6] [ 7] [ 8] [ 9] [ 10] [ 11] [ 12] [ 13] [ 14] [ 15] [ 16] [ 17] [ 18] [ 19]


Journal of Information Science and Engineering, Vol. 30 No. 4, pp. 1035-1052 (July 2014)

Tracking Pedestrian with Incrementally Learned Representation and Classification Model*

Beijing Lab of Intelligent Information
School of Computer Science
Beijing Institute of Technology
Beijing, 100081 P.R. China

Most of the existing tracking algorithms are challenged for the deficiency of handling non-stationary target appearance such as the drastic scale and perspective change of a moving pedestrian in the PTZ surveillance record. We propose a novel pedestrian tracking algorithm to cope with this problem by integrating incrementally learned representation and classification model. In the representation model, besides the widely used intensity template, the contour template with several sets of profiles from different perspectives is also employed to cope with the change of pedestrian contour. Both templates are updated incrementally during the tracking process to deal with the non-stationary appearance of the pedestrian. In the classification model, a multiple instance classier based on an incremental support vector machine is trained on-line as new observation becomes available. The learned classifier keeps the evolving representation model from drifting and enables reinitialization of the tracker once a failure occurs in the tracking process. The effectiveness of our algorithm is tested over several surveillance records captured from PTZ. The experiment results show that our algorithm can track the pedestrian more robustly than the other two compared cutting edge tracking algorithms.

Keywords: Pedestrian tracking, intensity and contour template, incremental multiple instance learning, PTZ visual surveillance, incremental principal components analysis

Full Text () Retrieve PDF document (201407_06.pdf)

Received June 27, 2012; revised October 6, 2012; accepted December 4, 2012.
Communicated by Chung-Lin Huang.
* This work was supported by National Nature Science Foundation of China (No. 61203291 and 90920009), NSFC-Guangdong Joint Fund (No. U1035004), and Specialized Fund for Joint Building Program of Beijing Municipal Education Commission.