Skip to content

Code of ICMI 2021 paper: "Predicting Gaze from Egocentric Social Interaction Videos and IMU Data"

Notifications You must be signed in to change notification settings

IIT-PAVIS/MultimodalGaze

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MultimodalGaze

Code of ICMI 2021 paper: "Predicting Gaze from Egocentric Social Interaction Videos and IMU Data"

You need to download additional modules to run the repo.

Flownet [1]

git clone https://github.com/ClementPinard/FlowNetPytorch.git

I3d network [2]

https://github.com/piergiaj/pytorch-i3d.git 

To run the model :

python3 main.py --model <model_type> 

(model types : 0 - signal, 1 - vision, 2 - multimodal)

To specify which multimodal & vision based model you need, you can change it in main.py file on line no. 29 for vision model (optical flow / i3d), and line no. 31

Select model(s) from models.py

i3d models:

Vision only : VISION_PIPELINE, Multimodal: FUSION_PIPELINE

flownet models:

Vision only : Flownet_PIPELINE, Multimodal: OF_FUSION_PIPELINE

Data

The models were trained on a new dataset collected by the authors for Social Interaction based scenarios in the wild. It will be made public in future.

All the checkpoints will be made available when the dataset will be made public.

References

[1] Dosovitskiy, et al. ICCV (2015). FlowNet: Learning Optical Flow with Convolutional Networks : http://lmb.informatik.uni-freiburg.de/Publications/2015/DFIB15

[2] Carreira, et al. CVPR (2017). Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset : https://openaccess.thecvf.com/content_cvpr_2017/html/Carreira_Quo_Vadis_Action_CVPR_2017_paper.html

About

Code of ICMI 2021 paper: "Predicting Gaze from Egocentric Social Interaction Videos and IMU Data"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages