Replies: 1 comment 3 replies
-
For Mixtral you will have to add the lora weights to each expert separately and they have different names. The code I sent above is meant for Mistral / Lama style models. For example:
Should work for the first weight of the |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
As per this discussion, I managed to get rank to 128 and QKVO layers on (yay!) but unfortunately no luck on the mlp weights yet. Here's the code I'm using for the lora utils.py tuning file in mlx-lm:
When I un-comment the lines about mlp, I get this error:
Am I listing the mlp weights wrong?
Beta Was this translation helpful? Give feedback.
All reactions