-
Notifications
You must be signed in to change notification settings - Fork 17
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Some questions and confirmation information #1
Comments
同问,目前不好复现, |
Hi ZR, Thanks for your interest!
Regarding the video length, we apologize that the current open-source version only supports videos shorter than 10 seconds. The timeline for open-sourcing additional versions, including the I2V model, is still undecided. |
@zRzRzRzRzRzRzR could you kindly share your full steps from start to end as how you created this video? How did you download checkpoints , what library versions did you use etc as I am trying to follow their READ ME but its incomplete. I am getting following error: from models.VchitectXL import VchitectXLTransformerModel |
Hi, |
I wll upload later |
Step 1
Step 2, modify the code
Change to
Step 3, if you want to change the number of frames of the video generation length:
Step 4, run the program:
This will run, I believe it will help you. |
|
Dear Development Team,
Hello, I have successfully installed the model and run it according to the requirements in the README, but I have encountered some issues and look forward to your response.
negative_prompt
In this case, I am unsure how to structure the prompt, so I simply wrote a prompt:
and set the seed to 42:
I set the output to 720x480 according to the README, and configured it as follows:
It occupied 67904MiB of GPU memory. The other parameters remained unchanged, with 50 sampling steps. The final video can be found here:
sample_1_seed0.mp4
Is this the expected result?
I did not see any relevant details about I2V in the code, nor any place where an image can be used as input. Should I understand that this open-source model is a T2V model?
It seems that there is no parameter to control the frame rate.
However, the video I generated only has 8 frames, with a total of 40 frames, as verified using the following command:
Is it because the open-source model only outputs 8 frames?
Additionally, there may be some issues in the code within the repository:
Vchitect-2.0/models/pipeline.py
Line 198 in 0ef47a5
This should be modified to
device = "cuda"
, or adddevice = "cuda"
in:Vchitect-2.0/inference.py
Line 15 in 0ef47a5
Otherwise, a tensor not on the same device error will occur during pos embed.
https://github.com/Vchitect/Vchitect-2.0/tree/master/models/__pycache__
Should this be deleted? It seems unnecessary.
Looking forward to your response.
The text was updated successfully, but these errors were encountered: