This repository contains the code used to produce the baseline results (TA3N) for the EPIC-KITCHENS-100 UDA Challenge.
Some modifications have been made to the original TA3N code base to produce results on EPIC-KITCHENS-100 including:
- Multiple classification heads to produce predictions for verb and nouns.
- Modified dataloader to load EPIC-KITCHENS-100 pre-extracted features.
- Modified training scripts to run the EPIC-KITCHENS-100 UDA baselines based. These have been modified from
script_train_val.sh
in the original repository.
The original TA3N code and ICCV publication can be found in the provided hyperlinks.
We provide modified training scripts for TA3N to replicate EPIC-100 UDA Challenge results.
-
Install dependecies
conda env create -f environment.yml
-
Download annotations and features.
- Annotations should be downloaded to the folder
./annotations
and extracted features to./data
. - Alteratively variables in the bash scripts (
path_labels_root="annotations"
,path_data_root="data
) can be modified to match the location of the labels/features on your machine.
- Annotations should be downloaded to the folder
-
Replicate results with the provided bash scripts:
./script_test_ta3n.sh
to re-produce TA3N domain adaptation results./script_test_source_only.sh
to re-produce source-only results
-
Each script will generate a submission script
test.json
which can be uploaded to the UDA Codalab challenge to replicate the results. To re-train models, settrain=true
in the bash scripts.
For futher details please see original source code for Temporal Attentive Alignment: https://github.com/cmhungsteve/TA3N
If you find this repository useful, please cite both the EPIC-KITCHENS dataset papers and the TA3N authors publications:
- EPIC-KITCHENS
@article{Damen2020RESCALING,
title={Rescaling Egocentric Vision},
author={Damen, Dima and Doughty, Hazel and Farinella, Giovanni Maria and and Furnari, Antonino
and Ma, Jian and Kazakos, Evangelos and Moltisanti, Davide and Munro, Jonathan
and Perrett, Toby and Price, Will and Wray, Michael},
journal = {CoRR},
volume = {abs/2006.13256},
year = {2020},
ee = {http://arxiv.org/abs/2006.13256},
}
@INPROCEEDINGS{Damen2018EPICKITCHENS,
title={Scaling Egocentric Vision: The EPIC-KITCHENS Dataset},
author={Damen, Dima and Doughty, Hazel and Farinella, Giovanni Maria and Fidler, Sanja and
Furnari, Antonino and Kazakos, Evangelos and Moltisanti, Davide and Munro, Jonathan
and Perrett, Toby and Price, Will and Wray, Michael},
booktitle={European Conference on Computer Vision (ECCV)},
year={2018}
}
- Temporal Attentive Alignment
@article{chen2019taaan,
title={Temporal Attentive Alignment for Large-Scale Video Domain Adaptation},
author={Chen, Min-Hung and Kira, Zsolt and AlRegib, Ghassan and Yoo, Jaekwon and Chen, Ruxin and Zheng, Jian},
booktitle = {International Conference on Computer Vision (ICCV)},
year={2019},
url={https://arxiv.org/abs/1907.12743}
}
@article{chen2019temporal,
title={Temporal Attentive Alignment for Video Domain Adaptation},
author={Chen, Min-Hung and Kira, Zsolt and AlRegib, Ghassan},
booktitle = {CVPR Workshop on Learning from Unlabeled Videos},
year={2019},
url={https://arxiv.org/abs/1905.10861}
}