From c9fe32761ec8d02eb80743a850000efca07fa4f3 Mon Sep 17 00:00:00 2001 From: NanoCode012 <kevinvong@rocketmail.com> Date: Sat, 16 Mar 2024 14:26:28 +0900 Subject: [PATCH] chore: lint --- src/axolotl/utils/models.py | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/src/axolotl/utils/models.py b/src/axolotl/utils/models.py index 29461e0e4d..40090a07c0 100644 --- a/src/axolotl/utils/models.py +++ b/src/axolotl/utils/models.py @@ -888,7 +888,9 @@ def load_and_quantize_parallel(name_param, model, **kwargs): if cfg.adapter in ["lora", "qlora"]: if cfg.gradient_checkpointing: - model.gradient_checkpointing_enable(gradient_checkpointing_kwargs=cfg.gradient_checkpointing_kwargs) + model.gradient_checkpointing_enable( + gradient_checkpointing_kwargs=cfg.gradient_checkpointing_kwargs + ) if ( cfg.load_in_8bit or cfg.load_in_4bit ) and not skip_prepare_model_for_kbit_training: