Skip to content

Latest commit

 

History

History
83 lines (74 loc) · 3.86 KB

File metadata and controls

83 lines (74 loc) · 3.86 KB

Leveraging Unimodal Self-Supervised Learning for Multimodal Audio-Visual Speech Recognition

This is the official PyTorch implementation of paper Leveraging Unimodal Self-supervised Learning for Multimodal Audio-visual Speech Recognition

PWC

PWC

PWC

PWC

Install the environment

  1. Clone the repo into a directory.
git clone https://github.com/LUMIA-Group/Leveraging-Self-Supervised-Learning-for-AVSR.git
  1. Install all required packages.
pip install -r requirements.txt

Noted that the Pytorch-Lightning lib do not support a wrapped ReduceLROnPlateau scheduler, we need to modify the lib manually by:

python -c "exec(\"import pytorch_lightning\nprint(pytorch_lightning.__file__)\")"
vi /path/to/pytorch_lightning/trainer/optimizers.py

and comments the 154-156 lines

# scheduler["reduce_on_plateau"] = isinstance(
#     scheduler["scheduler"], optim.lr_scheduler.ReduceLROnPlateau
# )

Preprocess the dataset

  1. Download LRW dataset and LRS2 dataset
  2. Download pretrained MoCo v2 model and wav2vec 2.0 model
  3. Change the directory in config.py to "relative directory" relative to the project root directory
  4. Preprocessing the LRW dataset.
cd trainFrontend
python saveh5.py
  1. Preprocessing the LRS2 dataset.
python saveh5.py

Training

  1. Train the visual front-end on LRW.
python trainfrontend.py
  1. Change the args["MOCO_FRONTEND_FILE"] in config.py to the trained front-end file, and config args["MODAL"] to choose modality.
  2. Train the AO and VO model first.
python train.py
  1. Then train the AV model. Before that, change the args["TRAINED_AO_FILE"] and args["TRAINED_VO_FILE"] to the trained AO and VO model.
python train.py

Evaluation

  1. Choose test configuration and model.
  2. Evaluate the visual word classification performance.
python evalfrontend.py
  1. Evaluate the AO/VO/AV model.
python eval.py

Cite

If you find this repo useful in your research, please consider citing it in the following format:

@inproceedings{pan2022leveraging,
  title={Leveraging Unimodal Self-Supervised Learning for Multimodal Audio-Visual Speech Recognition},
  author={Pan, Xichen and Chen, Peiyu and Gong, Yichen and Zhou, Helong and Wang, Xinbing and Lin, Zhouhan},
  booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
  pages={4491--4503},
  year={2022}
}