Replies: 3 comments
-
I also had this problem, but I tried to lower network_dim, or lower the learning rate,It's working.! |
Beta Was this translation helpful? Give feedback.
0 replies
-
-optimizer_type="AdamW8bit" change AdamW |
Beta Was this translation helpful? Give feedback.
0 replies
-
Thanks, might have been network dim, LR or optimizer type didn't have an effect. I ended up installing Windows and it's working without any issue. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Not able to get very far into training. I've been attempting to train LoRAs (standard), however the loss becomes nan, sometimes infinite regardless of learning rate. After around 300 steps the loss continues to increase and shortly after gets to nan, garbage samples are produced. e.g. training with 20 to 500 512x512 images & SD 1.5 model. Using an Nvidia 4090, fresh install of Ubuntu 22.04.1. Tried with fp16, bp16 and with/without xformers. Pulled the latest and re-ran setup. e.g. command:
accelerate launch --num_cpu_threads_per_process=2 "train_network.py" --enable_bucket --pretrained_model_name_or_path="runwayml/stable-diffusion-v1-5" --train_data_dir="/home/user/landscape/img" --resolution=512,512 --output_dir="/home/user/landscape/output" --logging_dir="/home/user/landscape/logs" --network_alpha="128" --save_model_as=safetensors --network_module=networks.lora --text_encoder_lr=5e-05 --unet_lr=0.001 --network_dim=128 --output_name="landscape01" --lr_scheduler_num_cycles="1" --learning_rate="0.00001" --lr_scheduler="constant" --train_batch_size="2" --max_train_steps="130" --save_every_n_epochs="1" --mixed_precision="fp16" --save_precision="fp16" --seed="1234" --caption_extension=".txt" --cache_latents --optimizer_type="AdamW8bit" --max_data_loader_n_workers="1" --clip_skip=2 --bucket_reso_steps=64 --xformers --bucket_no_upscale
I haven't been able to find anything regarding this issue elsewhere.
Beta Was this translation helpful? Give feedback.
All reactions