TY - GEN
T1 - 3DFaceMAE
T2 - 7th Chinese Conference on Pattern Recognition and Computer Vision, PRCV 2024
AU - Gao, Ziqi
AU - Li, Qiufu
AU - Shen, Linlin
AU - Yang, Junpeng
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2025.
PY - 2025
Y1 - 2025
N2 - Compared to 2D face recognition, 3D face recognition exhibits stronger robustness against variations like pose and illumination. However, due to the limited training data, the accuracy of existing 3D face recognition methods is still unsatisfactory. In this paper, we introduce 3DFaceMAE, which is the first masked autoencoder (MAE) based 3D face recognition method using point clouds. Specifically, we first synthesize a large-scale 3D point cloud facial dataset and combine it with the small-scale real data. In the pre-training of 3DFaceMAE, we extract the key facial regions from the input 3D facial point cloud, using normal difference techniques, and reconstruct these key regions using patch-based random masking reconstruction and super-resolution. We finally fine-tune the encoder of 3DFaceMAE on the real 3D face point cloud data. In the experiments, we test 3DFaceMAE on three 3D face datasets, as high as 91.17% was achieved on the Lock3DFace dataset, which is the first reported result surpassing 90%. In addition, the experimental results indicate that 3DFaceMAE has strong cross-quality generalization performance. We also validate the effectiveness of different components of 3DFaceMAE through ablation study.
AB - Compared to 2D face recognition, 3D face recognition exhibits stronger robustness against variations like pose and illumination. However, due to the limited training data, the accuracy of existing 3D face recognition methods is still unsatisfactory. In this paper, we introduce 3DFaceMAE, which is the first masked autoencoder (MAE) based 3D face recognition method using point clouds. Specifically, we first synthesize a large-scale 3D point cloud facial dataset and combine it with the small-scale real data. In the pre-training of 3DFaceMAE, we extract the key facial regions from the input 3D facial point cloud, using normal difference techniques, and reconstruct these key regions using patch-based random masking reconstruction and super-resolution. We finally fine-tune the encoder of 3DFaceMAE on the real 3D face point cloud data. In the experiments, we test 3DFaceMAE on three 3D face datasets, as high as 91.17% was achieved on the Lock3DFace dataset, which is the first reported result surpassing 90%. In addition, the experimental results indicate that 3DFaceMAE has strong cross-quality generalization performance. We also validate the effectiveness of different components of 3DFaceMAE through ablation study.
KW - 3D point cloud
KW - Face recognition
UR - http://www.scopus.com/inward/record.url?scp=85209347536&partnerID=8YFLogxK
U2 - 10.1007/978-981-97-8795-1_33
DO - 10.1007/978-981-97-8795-1_33
M3 - Conference contribution
AN - SCOPUS:85209347536
SN - 9789819787944
T3 - Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
SP - 488
EP - 503
BT - Pattern Recognition and Computer Vision - 7th Chinese Conference, PRCV 2024, Proceedings
A2 - Lin, Zhouchen
A2 - Zha, Hongbin
A2 - Cheng, Ming-Ming
A2 - He, Ran
A2 - Liu, Cheng-Lin
A2 - Ubul, Kurban
A2 - Silamu, Wushouer
A2 - Zhou, Jie
PB - Springer Science and Business Media Deutschland GmbH
Y2 - 18 October 2024 through 20 October 2024
ER -