Enhanced Normalized Mean Error loss for Robust Facial Landmark detection

Shenqi Lai (MeituanDianping Group), Zhenhua Chai (MeituanDianping Group), Huanhuan Meng (MeituanDianping Group), Shengxi Li (MeituanDianping Group), Mengzhao Yang (MeituanDianping Group), Xiaoming Wei (MeituanDianping Group)

Abstract
Normalized Mean Error (NME) is one of the most popular evaluation metrics in facial landmark detection benchmark. However, the commonly used loss functions (L1 and L2) are not designed to optimize NME directly, and thus there might be a gap between optimizing the distance losses for regressing the parameters of landmark coordinates and minimizing this metric value. In this paper, we will try to address this issue, and propose a novel loss function named Enhanced Normalized Mean Error (ENME) loss, which will consider both the final metric and the attention mechanism for different NME intervals. In order to evaluate the effectiveness of our proposed loss, we design and train a light-weight regressing model we call Thin Residual Network (TRNet). Extensive experiments are conducted on three popular public datasets such as AFLW, COFW and challenging 300W, and the results show that TRNet when trained with the enhanced NME loss will exhibit better performance than the state of the art methods.

DOI
10.5244/C.33.154
https://dx.doi.org/10.5244/C.33.154

Files
Paper (PDF)
Supplementary material (ZIP)

BibTeX
@inproceedings{BMVC2019,
title={Enhanced Normalized Mean Error loss for Robust Facial Landmark detection},
author={Shenqi Lai and Zhenhua Chai and Huanhuan Meng and Shengxi Li and Mengzhao Yang and Xiaoming Wei},
year={2019},
month={September},
pages={154.1--154.11},
articleno={154},
numpages={11},
booktitle={Proceedings of the British Machine Vision Conference (BMVC)},
publisher={BMVA Press},
editor={Kirill Sidorov and Yulia Hicks},
doi={10.5244/C.33.154},
url={https://dx.doi.org/10.5244/C.33.154}
}