Head pose estimation using deep multitask learning

2019 
Head pose estimation (HPE) plays a vital role in the field of human–computer interaction and remains a challenging task due to individual differences. To mitigate this issue, we propose a simple yet effective deep multitask learning framework for joint HPE and face verification (FV), where FV acts as an auxiliary task to boost the performance of HPE. The framework comprises a backbone net, data separate module, and two branches for HPE and FV, respectively. Considering other regions beyond those of the face provides useful information for HPE while FV should focus only on face regions, two kinds of regions, head and face, which share common feature representations in the backbone net, are separated by the data separate module and then fed into the appropriate branches. Kullback–Leibler divergence loss and L2-constrained softmax loss are connected to the end of the HPE branch and FV branch, respectively, to optimize the architecture. The proposed method is validated on three publicly available datasets: Pointing04, CAS-PEAL-R1, and CMU multi-PIE. The experimental results demonstrate that our method surpasses the state of the art, showing up to a 16.38% improvement on the famous benchmark, Pointing04. The best accuracies we report on these datasets are 89.53%, 99.74%, and 99.72%, respectively.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    49
    References
    1
    Citations
    NaN
    KQI
    []