The following table shows the comparison with MLLMs on the FaceXBench benchmark. The best performing model in each category is emboldened and the best model amongst all MLLMs is in purple.
The performance of FaceLLM models (FaceLLM-1B, FaceLLM-8B, and FaceLLM-38B) are compared in the following figures for different sub-tasks, including age estimation, gender prediction, race estimation, high-resolution face recognition, low-resolution face recognition, celebrity identification, face anti-spoofing, deepfake detection, attributes prediction, facial expression recognition, headpose estimation, face localization crowd counting, face parsing, and face tools retrieval.
The source code of our experiments as well as pretrained FaceLLM models are publicly available. We also release FairFaceGPT dataset.
@inproceedings{facellm2025,
title={Facellm: A multimodal large language model for face understanding},
author={Shahreza, Hatef Otroshi and Marcel, S{\'e}bastien},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={3677--3687},
year={2025}
}