torch.compile
not working with device_map
and multiple GPUs
#2592
Labels
torch.compile
not working with device_map
and multiple GPUs
#2592
System Info
Information
Tasks
no_trainer
script in theexamples
folder of thetransformers
repo (such asrun_no_trainer_glue.py
)Reproduction
I am unable to use
torch.compile
withdevice_map
when using multiple GPUs. I believe this issue is related to #2387.This can be reproduced with the following (the model does not seem to matter):
Stack trace
I have verified that this issue does not occur if I use
.cuda()
or enforce that the model run on a single GPU.Expected behavior
I expected the compilation to work with a single or with multiple GPUs.
The text was updated successfully, but these errors were encountered: