Skip to content

Commit

Permalink
minor typo fix (huggingface#33784)
Browse files Browse the repository at this point in the history
fix typo
  • Loading branch information
a-r-r-o-w authored Sep 30, 2024
1 parent b77846a commit 90dca5a
Show file tree
Hide file tree
Showing 2 changed files with 2 additions and 2 deletions.
2 changes: 1 addition & 1 deletion src/transformers/modeling_utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -2935,7 +2935,7 @@ def cuda(self, *args, **kwargs):

@wraps(torch.nn.Module.to)
def to(self, *args, **kwargs):
# For BNB/GPTQ models, we prevent users from casting the model to another dytpe to restrict unwanted behaviours.
# For BNB/GPTQ models, we prevent users from casting the model to another dtype to restrict unwanted behaviours.
# the correct API should be to load the model with the desired dtype directly through `from_pretrained`.
dtype_present_in_args = "dtype" in kwargs

Expand Down
2 changes: 1 addition & 1 deletion src/transformers/models/esm/modeling_esmfold.py
Original file line number Diff line number Diff line change
Expand Up @@ -1310,7 +1310,7 @@ def __init__(self, config):
def forward(self, residue_index, mask=None):
"""
Input:
residue_index: B x L tensor of indices (dytpe=torch.long) mask: B x L tensor of booleans
residue_index: B x L tensor of indices (dtype=torch.long) mask: B x L tensor of booleans
Output:
pairwise_state: B x L x L x pairwise_state_dim tensor of embeddings
Expand Down

0 comments on commit 90dca5a

Please sign in to comment.