Skip to content

Commit

Permalink
Remove learning rate scheduler in deepspeed config to avoid conflict (#…
Browse files Browse the repository at this point in the history
  • Loading branch information
Haoxiang-Wang authored Dec 4, 2023
1 parent 3e3229e commit 476a205
Show file tree
Hide file tree
Showing 3 changed files with 0 additions and 30 deletions.
10 changes: 0 additions & 10 deletions deepspeed/zero1.json
Original file line number Diff line number Diff line change
Expand Up @@ -24,16 +24,6 @@
"weight_decay": "auto"
}
},
"scheduler": {
"type": "WarmupDecayLR",
"params": {
"warmup_min_lr": "auto",
"warmup_max_lr": "auto",
"warmup_num_steps": "auto",
"warmup_type": "linear",
"total_num_steps": "auto"
}
},
"gradient_accumulation_steps": "auto",
"train_batch_size": "auto",
"train_micro_batch_size_per_gpu": "auto",
Expand Down
10 changes: 0 additions & 10 deletions deepspeed/zero2.json
Original file line number Diff line number Diff line change
Expand Up @@ -28,16 +28,6 @@
"weight_decay": "auto"
}
},
"scheduler": {
"type": "WarmupDecayLR",
"params": {
"warmup_min_lr": "auto",
"warmup_max_lr": "auto",
"warmup_num_steps": "auto",
"warmup_type": "linear",
"total_num_steps": "auto"
}
},
"gradient_accumulation_steps": "auto",
"train_batch_size": "auto",
"train_micro_batch_size_per_gpu": "auto",
Expand Down
10 changes: 0 additions & 10 deletions deepspeed/zero3.json
Original file line number Diff line number Diff line change
Expand Up @@ -32,16 +32,6 @@
"weight_decay": "auto"
}
},
"scheduler": {
"type": "WarmupDecayLR",
"params": {
"warmup_min_lr": "auto",
"warmup_max_lr": "auto",
"warmup_num_steps": "auto",
"warmup_type": "linear",
"total_num_steps": "auto"
}
},
"gradient_accumulation_steps": "auto",
"train_batch_size": "auto",
"train_micro_batch_size_per_gpu": "auto",
Expand Down

0 comments on commit 476a205

Please sign in to comment.