You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
However, on htop I see that only a single thread is being used. Even if I set torch.set_num_threads(32). It is extremely slow, making smoothquant unusable in my case.
I have a system with Intel® Xeon® Gold 5218 Processor.
Am I missing something? Thanks!
The text was updated successfully, but these errors were encountered:
Hi @akhauriyash , I was not able to reproduce this issue on several machines yet. Could you please share your enviroment where the issue occurs using pip list?
Hello,
I am trying to run the following script:
https://github.com/intel/neural-compressor/tree/master/examples/pytorch/nlp/huggingface_models/language-modeling/quantization/llm
I use the script below:
However, on htop I see that only a single thread is being used. Even if I set torch.set_num_threads(32). It is extremely slow, making smoothquant unusable in my case.
I have a system with Intel® Xeon® Gold 5218 Processor.
Am I missing something? Thanks!
The text was updated successfully, but these errors were encountered: