Ioannis Marras
Online learning and fusion of orientation appearance models for robust rigid object tracking
Marras, Ioannis; Tzimiropoulos, Georgios; Zafeiriou, Stefanos; Pantic, Maja
Authors
Georgios Tzimiropoulos
Stefanos Zafeiriou
Maja Pantic
Abstract
We introduce a robust framework for learning and fusing of orientation appearance models based on both texture and depth information for rigid object tracking. Our framework fuses data obtained from a standard visual camera and dense depth maps obtained by low-cost consumer depth cameras such as the Kinect. To combine these two completely different modalities, we propose to use features that do not depend on the data representation: angles. More specifically, our framework combines image gradient orientations as extracted from intensity images with the directions of surface normals computed from dense depth fields. We propose to capture the correlations between the obtained orientation appearance models using a fusion approach motivated by the original Active Appearance Models (AAMs). To incorporate these features in a learning framework, we use a robust kernel based on the Euler representation of angles which does not require off-line training, and can be efficiently implemented online. The robustness of learning from orientation appearance models is presented both theoretically and experimentally in this work. This kernel enables us to cope with gross measurement errors, missing data as well as other typical problems such as illumination changes and occlusions. By combining the proposed models with a particle filter, the proposed framework was used for performing 2D plus 3D rigid object tracking, achieving robust performance in very difficult tracking scenarios including extreme pose variations.
Citation
Marras, I., Tzimiropoulos, G., Zafeiriou, S., & Pantic, M. (2014). Online learning and fusion of orientation appearance models for robust rigid object tracking. Image and Vision Computing, 32(10), 707-727. https://doi.org/10.1016/j.imavis.2014.04.017
Journal Article Type | Article |
---|---|
Acceptance Date | Apr 5, 2014 |
Online Publication Date | May 20, 2014 |
Publication Date | Oct 14, 2014 |
Deposit Date | Jan 29, 2016 |
Publicly Available Date | Mar 29, 2024 |
Journal | Image and Vision Computing |
Print ISSN | 0262-8856 |
Electronic ISSN | 0262-8856 |
Publisher | Elsevier |
Peer Reviewed | Peer Reviewed |
Volume | 32 |
Issue | 10 |
Pages | 707-727 |
DOI | https://doi.org/10.1016/j.imavis.2014.04.017 |
Keywords | Rigid object tracking; Fusion of orientation appearance models; Subspace learning; Online learning; Face analysis; RGB-D |
Public URL | https://nottingham-repository.worktribe.com/output/738296 |
Publisher URL | http://www.sciencedirect.com/science/article/pii/S0262885614000924 |
Files
tzimiroIVC14.pdf
(6.7 Mb)
PDF
You might also like
Object landmark discovery through unsupervised adaptation
(2019)
Journal Article
Artificial intelligence-enhanced multi-material form measurement for additive materials
(2018)
Presentation / Conference
Deep word embeddings for visual speech recognition
(2018)
Conference Proceeding
End-to-end audiovisual speech recognition
(2018)
Conference Proceeding
Synergy between face alignment and tracking via Discriminative Global Consensus Optimization
(2017)
Conference Proceeding
Downloadable Citations
About Repository@Nottingham
Administrator e-mail: digital-library-support@nottingham.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search