This repository has been archived by the owner on Oct 9, 2023. It is now read-only.
How to use different video classifier backbones #1468
Unanswered
BernhardGlueck
asked this question in
Data / pipelines
Replies: 1 comment
-
Hi @BernhardGlueck Thanks for reporting this! This looks like a bug on our side (partially related to #1328) for the slowfast backends. Regarding the x3d backends, my guess is that they expect the image to be a bit bigger. Could you try creating your datamodule like this: from flash.video.classification.input_transform import VideoClassificationInputTransform
datamodule = VideoClassificationData.from_folders(
...,
transform=VideoClassificationInputTransform(image_size=512),
) That would set the image size to be 512 rather than the default of 244. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I am trying to train a video classifier using the following code:
My input data are thousands of video all with a resolution of 398x224 with 25 fps and a length of exactly 2 seconds each.
I tried different clip samples like random or uniform.
This works fine, but now i want to switch to different backbones.... in the example i am using x3d_xs from the tutorial.
However if i change that to any of the other supported backbones i get different errors:
What am i doing wrong ?
Beta Was this translation helpful? Give feedback.
All reactions