You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
My own task or dataset (give details below)
Reproduction (minimal, reproducible, runnable)
Reproduction:
fromtransformersimportAutoTokenizer, AutoModelForCausalLMfromoptimum.bettertransformerimportBetterTransformerimporttorchmodel_id="tiiuae/falcon-rw-1b"tokenizer=AutoTokenizer.from_pretrained(model_id)
model=AutoModelForCausalLM.from_pretrained(model_id)
model=BetterTransformer.transform(model)
inputs=tokenizer("Hello, my dog is cute", return_tensors="pt")
outputs=model.generate(**inputs, max_new_tokens=10)
Hi, Falcon with SDPA is supported by default in Transformers now huggingface/transformers#26572, and we deprecate the usage of BetterTransformer for this architecture.
System Info
Who can help?
@fxmarty
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction (minimal, reproducible, runnable)
Reproduction:
Falcon attention was refactored in huggingface/transformers@05ea7b7#diff-81c616a9db6f569c579ccf03c30c2f69aa7b65fa40959ac7e882fb8d541891d7. This removed the property
maybe_rotary
and adopted llama conventions for rotary embeddings.We could modify the use of
maybe_rotary
here by using something like:And then we'd need to adapt the code here, applying rotary embeddings when
alibi
is not in use.Expected behavior
Transformation would work.
The text was updated successfully, but these errors were encountered: