You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
ValueError: FP16 Mixed precision training with AMP or APEX (`--fp16`) and FP16 half precision evaluation (`--fp16_full_eval`) can only be used on CUDA devices.
So I tried adding --use_mps_device and --fp16 0 but got
File "/Users/usr/PycharmProjects/vigogne/finetune.py", line 200, in train
model = AutoModelForCausalLM.from_pretrained(
File "/Users/usr/PycharmProjects/vigogne/.venv/lib/python3.9/site-packages/transformers/models/auto/auto_factory.py", line 468, in from_pretrained
return model_class.from_pretrained(
File "/Users/usr/PycharmProjects/vigogne/.venv/lib/python3.9/site-packages/transformers/modeling_utils.py", line 2730, in from_pretrained
raise ValueError(
ValueError:
Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit
the quantized model. If you want to dispatch the model on the CPU or the disk while keeping
these modules in 32-bit, you need to set `load_in_8bit_fp32_cpu_offload=True` and pass a custom
`device_map` to `from_pretrained`. Check
https://huggingface.co/docs/transformers/main/en/main_classes/quantization#offload-between-cpu-and-gpu
for more details.
The text was updated successfully, but these errors were encountered:
But according to the last error message, you don't have enough memory to fit the model. You could try dispatching the model between CPU and GPU. See here
Hi, any idea if
finetune.py
should run on MacOS ?I tried the readme command
and got
So I tried adding
--use_mps_device
and--fp16 0
but gotThe text was updated successfully, but these errors were encountered: