Ioannis Marras
Online learning and fusion of orientation appearance models for robust rigid object tracking
Marras, Ioannis; Medina, Joan Alabort; Tzimiropoulos, Georgios; Zafeiriou, Stefanos; Pantic, Maja
Authors
Joan Alabort Medina
Georgios Tzimiropoulos
Stefanos Zafeiriou
Maja Pantic
Abstract
We present a robust framework for learning and fusing different modalities for rigid object tracking. Our method fuses data obtained from a standard visual camera and dense depth maps obtained by low-cost consumer depths cameras such as the Kinect. To combine these two completely different modalities, we propose to use features that do not depend on the data representation: angles. More specifically, our method combines image gradient orientations as extracted from intensity images with the directions of surface normal computed from dense depth fields provided by the Kinect. To incorporate these features in a learning framework, we use a robust kernel based on the Euler representation of angles. This kernel enables us to cope with gross measurement errors, missing data as well as typical problems in visual tracking such as illumination changes and occlusions. Additionally, the employed kernel can be efficiently implemented online. Finally, we propose to capture the correlations between the obtained orientation appearance models using a fusion approach motivated by the original AAM. Thus the proposed learning and fusing framework is robust, exact, computationally efficient and does not require off-line training. By combining the proposed models with a particle filter, the proposed tracking framework achieved robust performance in very difficult tracking scenarios including extreme pose variations.
Citation
Marras, I., Medina, J. A., Tzimiropoulos, G., Zafeiriou, S., & Pantic, M. (2013). Online learning and fusion of orientation appearance models for robust rigid object tracking. In 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG). https://doi.org/10.1109/FG.2013.6553798
Conference Name | 2013 10th IEEE International Conference on Automatic Face & Gesture Recognition and Workshops (FG 2013) |
---|---|
Start Date | Apr 22, 2013 |
End Date | Apr 26, 2013 |
Publication Date | Jul 15, 2013 |
Deposit Date | Jan 29, 2016 |
Publicly Available Date | Jan 29, 2016 |
Publisher | Institute of Electrical and Electronics Engineers |
Peer Reviewed | Peer Reviewed |
Book Title | 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (FG) |
ISBN | 978-1-4673-5545-2 |
DOI | https://doi.org/10.1109/FG.2013.6553798 |
Public URL | https://nottingham-repository.worktribe.com/output/1004912 |
Publisher URL | http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6553798 |
Additional Information | © 2013 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. |
Files
tzimiroFG13.pdf
(<nobr>1.6 Mb</nobr>)
PDF
You might also like
To Learn Image Super-Resolution, Use a GAN to Learn How to Do Image Degradation First
(2018)
Conference Proceeding
Zero-shot keyword spotting for visual speech recognition in-the-wild
(2018)
Conference Proceeding
Artificial intelligence-enhanced multi-material form measurement for additive materials
(2018)
Presentation / Conference
Object landmark discovery through unsupervised adaptation
(2019)
Journal Article
Hierarchical binary CNNs for landmark localization with limited resources
(2018)
Journal Article