Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

support for llama3.2 vision #30

Open
vaiju1981 opened this issue Dec 17, 2024 · 2 comments
Open

support for llama3.2 vision #30

vaiju1981 opened this issue Dec 17, 2024 · 2 comments

Comments

@vaiju1981
Copy link

First of all thanks for the amazing work. It helps us build a very simple yet efficient router within our Java applications.

I was wondering if there is any plan to support LLama3.2 vision models.

--Thanks and Regards
Vaijanath

@mukel
Copy link
Owner

mukel commented Dec 17, 2024

I looked into implementing the vision encoder component, specially for QwenVL models, which were merged into llama.cpp just a few days ago. I work on this on my spare time, which is not much lately. To make it easier in the future, I'm working on a simple tensor library for inference in Java. Slowly but I'm on it, I really enjoy hacking on this.

@vaiju1981
Copy link
Author

If you can provide with tensor library, I can take a stab at it. Right now in order to make llama3.2 vision to work with current code i need to make weights to have List<FloatTensor[]> and have identity operations for missing layers.

for example attn_q.weight is not available for all the layers.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants