One strategy to train lightweight and efficient face recognition networks is to train a face recognition network on the large-scale face recognition datasets. However, training an efficient face recognition model using large-scale face recognition datasets requires access to such a dataset. Nonetheless, large-scale face recognition datasets, such as MS-Celeb, etc., were collected by crawling images from the Internet, thus raising legal, ethical, and privacy concerns. To address such concerns, recently several works proposed generating synthetic face datasets and use the synthetic face images for training face recognition models. However, generating synthetic face datasets with sufficient inter-class and intra-class variations is still a challenging problem. Another strategy for training a lightweight face recognition model is to transfer the knowledge of a model trained on a large dataset to a lightweight network through knowledge distillation. Notwithstanding, the knowledge distillation from a teacher model often requires access to the original or another large-scale real dataset, which has several challenges. In this work, we propose a new framework, named SynthDistill, to distill the knowledge of a pretrained teacher using synthetic face images without identity labels, thus mitigating the need for real identity-labeled data during the distillation phase. We propose dynamic sampling from the intermediate latent space of a StyleGAN to generate new images and enhance training. Hence, our proposed knowledge distillation framework does not require real face images during training.
We use an online generation of synthetic images and train the lightweight network along with the image generation within a loop through knowledge distillation based framework. We use StyleGAN as a pretrained face generator network and deploy a dynamic sampling approach to generate synthetic face images though a feedback mechanism during training. We use the generated face images to train the lightweight network as a student within our knowledge distillation framework. Based on the teacher-student agreement, we dynamically re-sampled from the intermediate latent space of StyleGAN. For samples with low similarity between embeddings of teacher and student we re-sample similar latent code (to help training difficult samples), but for high similarity we re-sample different latent code (to help generalisation), enabling the training of more robust training. Compared to previous works for the training of face recognition models on synthetic datasets, our proposed knowledge distillation framework does not require identity labels in the training, simplifying the process of generating synthetic face images.
Given the effectiveness of our method, SynthDistill achieved the first rank in training face recognition with unlimited synthetic data (sub-task 2.2) of the FRCSyn challenge at CVPR 2024:
The source code and pretrained models are available in the following GitLab repository.
@article{access2024synthdistill,
title={Knowledge Distillation for Face Recognition using Synthetic Data with Dynamic Latent Sampling},
author={Shahreza, Hatef Otroshi and George, Anjith and Marcel, S{\'e}bastien},
journal={IEEE Access},
year={2024},
publisher={IEEE}
}
@inproceedings{ijcb2023synthdistill,
title={SynthDistill: Face recognition with knowledge distillation from synthetic data},
author={Shahreza, Hatef Otroshi and George, Anjith and Marcel, S{\'e}bastien},
booktitle={2023 IEEE International Joint Conference on Biometrics (IJCB)},
pages={1--10},
year={2023},
organization={IEEE}
}