Skip to main content

Research Repository

Advanced Search

Online learning and fusion of orientation appearance models for robust rigid object tracking

Marras, Ioannis; Medina, Joan Alabort; Tzimiropoulos, Georgios; Zafeiriou, Stefanos; Pantic, Maja

Online learning and fusion of orientation appearance models for robust rigid object tracking Thumbnail


Authors

Ioannis Marras

Joan Alabort Medina

Georgios Tzimiropoulos

Stefanos Zafeiriou

Maja Pantic



Abstract

We present a robust framework for learning and fusing different modalities for rigid object tracking. Our method fuses data obtained from a standard visual camera and dense depth maps obtained by low-cost consumer depths cameras such as the Kinect. To combine these two completely different modalities, we propose to use features that do not depend on the data representation: angles. More specifically, our method combines image gradient orientations as extracted from intensity images with the directions of surface normals computed from dense depth fields provided by the Kinect. To incorporate these features in a learning framework, we use a robust kernel based on the Euler representation of angles. This kernel enables us to cope with gross measurement errors, missing data as well as typical problems in visual tracking such as illumination changes and occlusions. Additionally, the employed kernel can be efficiently implemented online. Finally, we propose to capture the correlations between the obtained orientation appearance models using a fusion approach motivated by the original AAM. Thus the proposed learning and fusing framework is robust, exact, computationally efficient and does not require off-line training. By combining the proposed models with a particle filter, the proposed tracking framework achieved robust performance in very difficult tracking scenarios including extreme pose variations. © 2013 IEEE.

Citation

Marras, I., Medina, J. A., Tzimiropoulos, G., Zafeiriou, S., & Pantic, M. (2013, April). Online learning and fusion of orientation appearance models for robust rigid object tracking. Presented at 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, FG 2013, Shanghai, China

Presentation Conference Type Edited Proceedings
Conference Name 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, FG 2013
Start Date Apr 22, 2013
End Date Apr 26, 2013
Online Publication Date Jul 15, 2013
Publication Date Aug 20, 2013
Deposit Date Jan 29, 2016
Publicly Available Date Jan 29, 2016
Publisher Institute of Electrical and Electronics Engineers
Peer Reviewed Peer Reviewed
Pages 290-297
Book Title 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG)
ISBN 9781467355452
DOI https://doi.org/10.1109/FG.2013.6553798
Public URL https://nottingham-repository.worktribe.com/output/1004912
Publisher URL http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6553798
Additional Information © 2013 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.

Files





Downloadable Citations