You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,
While applying quantization on Llama model, we first convert weights downloaded from Meta and then use huggingface converter and then apply huggingface compatible AWQ quantization.
Is there a quantization tool specific to AMD, where the dependency on huggingface is removed?
Thanks,
Ashima
The text was updated successfully, but these errors were encountered:
Are you trying to convert PyTorch model to ONNX model? Then yes, today we use Hugging Face converter. I will check about the possible other option for future and update.
Hello,
While applying quantization on Llama model, we first convert weights downloaded from Meta and then use huggingface converter and then apply huggingface compatible AWQ quantization.
Is there a quantization tool specific to AMD, where the dependency on huggingface is removed?
Thanks,
Ashima
The text was updated successfully, but these errors were encountered: