TY - JOUR
T1 - Exploring transfer learning approaches for head pose classification from multi-view surveillance images
AU - Kolar Rajagopal, Anoop
AU - Subramanian, Ramanathan
AU - Ricci, Elisa
AU - Vieriu, Radu L.
AU - Lanz, Oswald
AU - Kalpathi R., Ramakrishnan
AU - Sebe, Nicu
N1 - Funding Information:
Acknowledgments The authors gratefully acknowledge partial support from Singapore’s Agency for Science, Technology and Research (A*STAR) under the Human Sixth Sense Programme (HSSP) grant, EIT ICT Labs SSP 12205 Activity TIK—The Interaction Toolkit, tasks T1320A-T1321A and the FP7 EU project DALI.
PY - 2014/8
Y1 - 2014/8
N2 - Head pose classification from surveillance images acquired with distant, large field-of-view cameras is difficult as faces are captured at low-resolution and have a blurred appearance. Domain adaptation approaches are useful for transferring knowledge from the training (source) to the test (target) data when they have different attributes, minimizing target data labeling efforts in the process. This paper examines the use of transfer learning for efficient multi-view head pose classification with minimal target training data under three challenging situations: (i) where the range of head poses in the source and target images is different, (ii) where source images capture a stationary person while target images capture a moving person whose facial appearance varies under motion due to changing perspective, scale and (iii) a combination of (i) and (ii). On the whole, the presented methods represent novel transfer learning solutions employed in the context of multi-view head pose classification. We demonstrate that the proposed solutions considerably outperform the state-of-the-art through extensive experimental validation. Finally, the DPOSE dataset compiled for benchmarking head pose classification performance with moving persons, and to aid behavioral understanding applications is presented in this work.
AB - Head pose classification from surveillance images acquired with distant, large field-of-view cameras is difficult as faces are captured at low-resolution and have a blurred appearance. Domain adaptation approaches are useful for transferring knowledge from the training (source) to the test (target) data when they have different attributes, minimizing target data labeling efforts in the process. This paper examines the use of transfer learning for efficient multi-view head pose classification with minimal target training data under three challenging situations: (i) where the range of head poses in the source and target images is different, (ii) where source images capture a stationary person while target images capture a moving person whose facial appearance varies under motion due to changing perspective, scale and (iii) a combination of (i) and (ii). On the whole, the presented methods represent novel transfer learning solutions employed in the context of multi-view head pose classification. We demonstrate that the proposed solutions considerably outperform the state-of-the-art through extensive experimental validation. Finally, the DPOSE dataset compiled for benchmarking head pose classification performance with moving persons, and to aid behavioral understanding applications is presented in this work.
KW - Moving persons
KW - Multi-view head pose classification
KW - Transfer learning
KW - Varying acquisition conditions
UR - http://www.scopus.com/inward/record.url?scp=84902284721&partnerID=8YFLogxK
U2 - 10.1007/s11263-013-0692-2
DO - 10.1007/s11263-013-0692-2
M3 - Article
AN - SCOPUS:84902284721
SN - 0920-5691
VL - 109
SP - 146
EP - 167
JO - International Journal of Computer Vision
JF - International Journal of Computer Vision
IS - 1-2
ER -