Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Scale decoding architectures to lower parameter counts and to fit on smaller GPUs #6

Open
reesekneeland opened this issue Apr 9, 2024 · 0 comments
Labels
limited-gpu Issues that can be addressed with access to limited GPUs (less than an A100)

Comments

@reesekneeland
Copy link
Collaborator

MindEye 1 and 2 in their default training/inference configurations require an A100 to use. There has been other recent work exploring a reduction in parameter counts that could be valuable to implement, in service of our tertiary goal of making these decoding algorithms more scalable and easier to use. This is also a good item for people with limited compute (no A100s) to work on.

Lite-Mind paper: https://arxiv.org/html/2312.03781v1

Other easy things:

  • Don't load all of the images onto the CPU
  • Smaller batch sizes
  • Disable unnecessary modules (captioning module, etc)
@reesekneeland reesekneeland added the limited-gpu Issues that can be addressed with access to limited GPUs (less than an A100) label Apr 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
limited-gpu Issues that can be addressed with access to limited GPUs (less than an A100)
Projects
None yet
Development

No branches or pull requests

1 participant