This repo is contained an implementation of the paper 2020 IEEE ICIP, Pairwise Adjacency Matrix on Spatial Temporal Graph Convolution Network for Skeleton-Based Two-Person Interaction Recognition(url: https://ieeexplore.ieee.org/document/9190680) and a maintained version of our original repo https://github.com/ajisetyoko/mutual-action
An extended version of ST-GCN [1] for Action Recognition focused on mutual action/Two-Person Interaction Recognition(TPIR).
- Pretrained for CV in PAM Mode
- Kinetics's Selected Action
- Kinetics Pretrained
All of experiment run on dependencies which lied on environment.yml. However, The original version on ST-GCN[1] use the requirements.txt as their dependencies.
- Original ST-GCN dependencies
pip install -r requirements.txt
cd torchlight; python setup.py install; cd ..
- Modified dependencies
pip install -r environment.yml
Example : to test model in MA Mode in CS_PP
main.py recognition -c config/MA_Mode/CS/pp.yaml
- NTURGB+D 120 [2]
- NTURGB+D 60 [2]
- Kinetics-Skeleton
- SBU-Kinect-Interaction dataset v2.0 [3]
- UT-Interaction Dataset [4]
We use extracted skeleton model which alredy done by [1]. This data can be downloaded from their repositories : GoogleDrive or BaiduYun
A. NTURGB+D 120
Model | Mode | CS | CV |
---|---|---|---|
PP Matrix | MA | 80.17 | 86.56 |
CP Matrix | MA | 78.93 | 82.87 |
PCP Matrix | MA | 83.28 | 88.36 |
PAM | MH | 82.1 | 80.91 |
PAM | AD | 73.87 | 76.85 |
PP=Pairwise of two partners; CP=partner-1 to the center of partner-2 and vice versa; PCP use both PP and CP; MA = trained and tested on mutual actions only. MH = Tested on mutual action subset only; AD=Tested on all actions label, *PCP
B. SBU-Kinect-Ineteraction
Test Model | FOLD 1 | FOLD2 | FOLD 3 | FOLD 4 | FOLD 5 | Average |
---|---|---|---|---|---|---|
TOP 1 Accuration | 93.10% | 98.53 | 94.21% | 91.47% | 98.53% | 94.6% |
TOP 5 Accuration | 100% | 100% | 100% | 100% | 100% | 100% |
C. UT-Interaction Dataset
- SET 1
Test Model | FOLD1 | FOLD2 | FOLD3 | FOLD4 | FOLD5 | Average |
---|---|---|---|---|---|---|
TOP 1 Accuration | 83.3% | 75.0% | 75.0% | 75.0% | 75.0% | 76.66% |
- SET 2
Test Model | FOLD1 | FOLD2 | FOLD3 | FOLD4 | FOLD5 | Average |
---|---|---|---|---|---|---|
TOP 1 Accuration | 83.3% | 75.0% | 75.0% | 83.3% | 66.7% | 77.30% |
- Combination
Test Model | FOLD1 | FOLD2 | FOLD3 | FOLD4 | FOLD5 | Average |
---|---|---|---|---|---|---|
TOP 1 Accuration | 100% | 83.3% | 83.3% | 100% | 83.3% | 90.1% |
[1] Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition, Sijie Yan and Yuanjun Xiong and Dahua Lin
[2] http://rose1.ntu.edu.sg/datasets/actionrecognition.asp
[3] https://www3.cs.stonybrook.edu/~kyun/research/kinect_interaction/index.html
[4] Semantic Description of Human Activities, http://cvrc.ece.utexas.edu/SDHA2010/Human_Interaction.html
[5] https://ieeexplore.ieee.org/document/9190680
@INPROCEEDINGS{9190680,
author={C. -L. {Yang} and A. {Setyoko} and H. {Tampubolon} and K. -L. {Hua}},
booktitle={2020 IEEE International Conference on Image Processing (ICIP)},
title={Pairwise Adjacency Matrix on Spatial Temporal Graph Convolution Network for Skeleton-Based Two-Person Interaction Recognition},
year={2020},
volume={},
number={},
pages={2166-2170},
doi={10.1109/ICIP40778.2020.9190680}}
For any question, feel free to contact us
Aji : [email protected]
Hendrik Tampubolon : [email protected]
Hendri Sutrisno : [email protected]