Skip to content

Commit

Permalink
try 4.35.2
Browse files Browse the repository at this point in the history
  • Loading branch information
dakinggg committed Dec 15, 2023
1 parent 6ab25b4 commit 291a91e
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 1 deletion.
1 change: 1 addition & 0 deletions llmfoundry/models/hf/hf_causal_lm.py
Original file line number Diff line number Diff line change
Expand Up @@ -109,6 +109,7 @@ def __init__(self, om_model_config: Union[DictConfig,
use_auth_token=use_auth_token,
# attn_implementation=requested_attention_implementation,
)
config._flash_attn_2_enabled = use_flash_attention_2

# This is not ideal, however Hugging Face's _autoset_attn_implementation function
# forces you to load the model in fp16/bf16 if you want to use flash attention. Rather than loading
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@
install_requires = [
'mosaicml[libcloud,wandb,mlflow,oci,gcs]>=0.17.1,<0.18',
'accelerate>=0.25,<0.26', # for HF inference `device_map`
'transformers>=4.36,<4.37',
'transformers>=4.35,<4.36',
'mosaicml-streaming>=0.7.1,<0.8',
'torch>=2.1,<2.1.1',
'datasets==2.15.0',
Expand Down

0 comments on commit 291a91e

Please sign in to comment.