From f44f0286be5e8fb3a5ff7428d8f48984d7f6dd51 Mon Sep 17 00:00:00 2001 From: xzuyn <16216325+xzuyn@users.noreply.github.com> Date: Fri, 5 Apr 2024 12:07:16 -0400 Subject: [PATCH] Remove `validate_quantized_dora` DoRA with quantized layers is supported with PEFT 0.10.0 --- .../utils/config/models/input/v0_4_1/__init__.py | 11 ----------- 1 file changed, 11 deletions(-) diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py index ad332da2da..464f2641a9 100644 --- a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py +++ b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py @@ -242,17 +242,6 @@ def validate_qlora(self): raise ValueError("Require cfg.load_in_4bit to be True for qlora") return self - @model_validator(mode="before") - @classmethod - def validate_quantized_dora(cls, data): - if data.get("peft_use_dora") and ( - data.get("load_in_8bit") or data.get("load_in_4bit") - ): - raise ValueError( - "`peft_use_dora` is not currently compatible with quantized weights." - ) - return data - class ReLoRAConfig(BaseModel): """ReLoRA configuration subset"""