-
Notifications
You must be signed in to change notification settings - Fork 848
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
OSError with Llama3.2-3B-Instruct-QLORA_INT4_EO8 - missing files? #194
Comments
The files we provide via I am curious what code needs these files and spitting out this error? |
Yes, I'm using transformers. I've tried using both transformers pipeline and automodel:
and automodel:
both methods produce the same error:
|
@StephenQuirolgico Can you kindly confirm what version of transformers you're using? |
transformers 4.46.0 |
Hey @StephenQuirolgico, we are working with HF to have these weights converted and supported in transformers. But for now, you can try either llama stack or export with ExecuTorch. Our official llama website has more detail on these. We might also better help you if you share which platform you plan to run inference on and what use cases you are trying to do. |
@WuhanMonkey I'm running this on RHEL 8. I have existing code using transformers and Llama3.2-3B, and just wanted to test the quantized version (by just swapping out the model in the code). Is there a rough timeframe on when these models will be supported in HF? |
When trying to run
Llama3.2-3B-Instruct-QLORA_INT4_EO8
, I'm getting the error:OSError: meta-llama/Llama-3.2-3B-Instruct-QLORA_INT4_EO8 does not appear to have a file named pytorch_model.bin, model.safetensors, tf_model.h5, model.ckpt or flax_model.msgpack.
I've tried using
transformers
to pull the model as well as downloading the model directly usingllama model download
. In both cases, the models are downloading successfully, so not sure why it is missing files.The text was updated successfully, but these errors were encountered: