This repository implements several basic data-augmentation transforms for pytorch video inputs
The idea was to produce the equivalent of torchvision transforms for video inputs. (The code is therefore widely based on the code from this repository :) )
The basic paradigm is that dataloading should produce videoclips as a list of PIL Images or numpy.ndarrays (in format as read by opencv).
Several transforms are then provided in video_transforms.
Each transform iterates on all the images in the list and applies the wanted augmentation.
We then have to convert those inputs to torch tensors. This can be produced by the volume_transform.ClipToTensor class, which produces a video volume in format (n_channels, n_images, height, width) where n_channels = 3 in case of images.
When randomness is involved, the same random parameters (crop size, scale size, rotation angle,...) are applied to all the frames.
Transforms can be composed just as in torchvision with video_transforms.Compose.
To quickly see a demo of the transformations, run python testtransforms.py
This should produce something like the top image (this is a dummy clip for now, so the same image is repeated several times)
- Compose
- Resize
- CenterCrop
- RandomCrop
- RandomHorizontalFlip
- RandomResize
- RandomRotation
- ColorJitter (acts on brightness, saturation, contrast and hue, only on PIL Images for now)
- RandomResizedCrop
-
ClipToTensor
-
Add profiling utilites
git clone https://github.com/hassony2/torch_videovision
export PYTHONPATH=$PYTHONPATH:path/to/torch_videovision
from videotranforms import video_transforms, volume_transforms
video_transform_list = [video_transforms.RandomRotation(30),
video_transforms.RandomCrop((200, 200))
volume_transforms.ClipToTensor()]
transforms = video_transforms.Compose(video_transform_list)