Fig. 1 Architecture of Multiscale Facial Expression Recognition based on Dynamic Global and Static Local Attention
- Download pre-trained model of MSCeleb.
- Download RAF-DB dataset and extract the
raf-basic
dir to./datasets
. - Download AffectNet dadtaset and extract the
AffectNet
dir to./datasets
. - Then
preprocess
the datasets as follow:
- We use the face alignment codes in face.evl to align face images first.
- the
aligned
face struct as follow:
- data/raf-db/
train/
train_00001_aligned.jpg # aligned by MTCNN
train_00002_aligned.jpg # aligned by MTCNN
...
valid/
test_0001_aligned.jpg # aligned by MTCNN
test_0002_aligned.jpg # aligned by MTCNN
...
the remain file or code (DSF Loss function and the pretrained weights of ConvNext model on Ms-Celeb-1M dataset will release after the paper was accepted !!!)
CUDA_VISIBLE_DEVICES=0 python train.py --help
Pre-trained models can be downloaded for evaluation as following:
dataset | accuracy | weight |
---|---|---|
RAF-DB |
92.08 |
Coming soon |
AffectNet-8 |
63.15 |
Coming soon |
AffectNet-7 |
67.06 |
Coming soon |
FERPlus |
91.09 |
Coming soon |
Our research code is released under the MIT license. See LICENSE for details.
And the remain file or code will release soon!!!!!!
you may want to cite:
@ARTICLE{10678884,
title={Multiscale Facial Expression Recognition Based on Dynamic Global and Static Local Attention},
author={Xu, Jie and Li, Yang and Yang, Guanci and He, Ling and Luo, Kexin},
journal={IEEE Transactions on Affective Computing},
year={2024},
volume={},
number={},
pages={1-14},
doi={10.1109/TAFFC.2024.3458464}}
Thanks for the code of the following:
ConvNext and
WZMIAOMIAO