diff --git a/requirements.txt b/requirements.txt index 456c63ca51..d100139ca4 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,9 +1,9 @@ --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/ packaging==23.2 -peft==0.13.2 +peft==0.14.0 transformers==4.46.3 tokenizers>=0.20.1 -bitsandbytes==0.44.1 +bitsandbytes==0.45.0 accelerate==1.1.0 datasets==3.1.0 deepspeed==0.15.4 diff --git a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py index c9170b7a84..24ea62c77f 100644 --- a/src/axolotl/utils/config/models/input/v0_4_1/__init__.py +++ b/src/axolotl/utils/config/models/input/v0_4_1/__init__.py @@ -1521,19 +1521,6 @@ def check_sample_packing_w_sdpa_bf16(cls, data): return data - @model_validator(mode="before") - @classmethod - def check_hopper_8bit_lora(cls, data): - is_sm_90: bool = ( - data["capabilities"] - and data["capabilities"].get("compute_capability") == "sm_90" - ) - if data.get("adapter") and data.get("load_in_8bit") and is_sm_90: - # see https://github.com/bitsandbytes-foundation/bitsandbytes/issues/538#issuecomment-2262945464 - raise ValueError("8-bit LoRA is not supported on Hopper GPUs") - - return data - @model_validator(mode="before") @classmethod def check_fsdp_deepspeed(cls, data): diff --git a/tests/e2e/multigpu/test_llama.py b/tests/e2e/multigpu/test_llama.py index d8dcf3118a..c9938a1559 100644 --- a/tests/e2e/multigpu/test_llama.py +++ b/tests/e2e/multigpu/test_llama.py @@ -14,8 +14,6 @@ from axolotl.utils.dict import DictDefault -from ..utils import is_hopper - LOG = logging.getLogger("axolotl.tests.e2e.multigpu") os.environ["WANDB_DISABLED"] = "true" @@ -144,7 +142,6 @@ def test_lora_ddp_packed(self, temp_dir, gradient_accumulation_steps): ] ) - @pytest.mark.skipif(is_hopper(), reason="h100 doesn't support 8-bit lora") def test_dpo_lora_ddp(self, temp_dir): # pylint: disable=duplicate-code cfg = DictDefault(