You need to download additional modules to run the repo.
Flownet [1]
git clone https://github.com/ClementPinard/FlowNetPytorch.git
I3d network [2]
https://github.com/piergiaj/pytorch-i3d.git
To run the model :
python3 main.py --model <model_type>
(model types : 0 - signal
, 1 - vision
, 2 - multimodal
)
To specify which multimodal & vision based model you need, you can change it in main.py file on line no. 29 for vision model (optical flow / i3d), and line no. 31
Select model(s) from models.py
i3d models:
Vision only : VISION_PIPELINE
, Multimodal: FUSION_PIPELINE
flownet models:
Vision only : Flownet_PIPELINE
, Multimodal: OF_FUSION_PIPELINE
The models were trained on a new dataset collected by the authors for Social Interaction based scenarios in the wild. It will be made public in future.
All the checkpoints will be made available when the dataset will be made public.
[1] Dosovitskiy, et al. ICCV (2015). FlowNet: Learning Optical Flow with Convolutional Networks : http://lmb.informatik.uni-freiburg.de/Publications/2015/DFIB15
[2] Carreira, et al. CVPR (2017). Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset : https://openaccess.thecvf.com/content_cvpr_2017/html/Carreira_Quo_Vadis_Action_CVPR_2017_paper.html