Skip to content

Commit

Permalink
disable future fix for now
Browse files Browse the repository at this point in the history
  • Loading branch information
winglian committed Dec 10, 2023
1 parent 99d68c4 commit 8335e40
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 1 deletion.
1 change: 1 addition & 0 deletions docker/Dockerfile-runpod
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ FROM winglian/axolotl:$BASE_TAG
ENV HF_DATASETS_CACHE="/workspace/data/huggingface-cache/datasets"
ENV HUGGINGFACE_HUB_CACHE="/workspace/data/huggingface-cache/hub"
ENV TRANSFORMERS_CACHE="/workspace/data/huggingface-cache/hub"
ENV HF_HOME="/workspace/data/huggingface-cache/hub"

COPY scripts/runpod-entrypoint.sh /root/runpod-entrypoint.sh

Expand Down
3 changes: 2 additions & 1 deletion src/axolotl/utils/models.py
Original file line number Diff line number Diff line change
Expand Up @@ -308,7 +308,8 @@ def load_model(
or cfg.is_falcon_derived_model
or cfg.is_mistral_derived_model
):
model_kwargs["attn_implementation"] = "flash_attention_2"
# TODO enable once properly supported in transformers
# model_kwargs["attn_implementation"] = "flash_attention_2"
model_kwargs["use_flash_attention_2"] = True # legacy, to be deprecated

try:
Expand Down

0 comments on commit 8335e40

Please sign in to comment.