-
Notifications
You must be signed in to change notification settings - Fork 82
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference with lower VRAM requirements #18
Conversation
Thanks a lot for your work along the time |
The batch inference is not changed by this PR. The only thing this PR change is adding some lines of torch.no_grad(). SanaPipeline class in the current repo only supports batch inference for the same prompt: Line 195 in c66ebf9
If you input a prompts list, it will generate it separately. |
then it is probably because of the changes i had to make it to run on windows - i probably broken something because i am giving only 1 prompt via gradio https://github.com/FurkanGozukara/Sana/blob/main/app/sana_pipeline.py |
your code has this for batch size
|
Yes. This is only for generating multiple images with a single prompt. If you want to batch inference with input like: ['car', 'car', 'car'], then it's not supported for now. |
I am using WSL2 and the sample script from the model's page on github. |
This PR breaks batch inference. Fixed in 9da8550. |
thanks i told batch was broken |
Inference with bfloat16 and float16 works on a RTX4070. Float32 gives OOM.