Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add embedding weight cast from float32 to bfloat16 #55

Merged
merged 1 commit into from
Dec 25, 2024

Conversation

dgolubovicTT
Copy link
Contributor

@dgolubovicTT dgolubovicTT commented Dec 25, 2024

Since ttnn only supports weight of embedding to be bfloat16, cast is introduced as a temporary solution until tt-mlir supports casting of embedding weight in workaround pass (tenstorrent/tt-mlir#1657). This enables llama embedding to work.

@dgolubovicTT dgolubovicTT self-assigned this Dec 25, 2024
…ch.py to cast weight to bfloat16 if they are float32.
@dgolubovicTT dgolubovicTT force-pushed the dgolubovic/add-embedding-weight-cast branch from 63fcdb6 to 0a429cf Compare December 25, 2024 15:32
Copy link
Contributor

@nvukobratTT nvukobratTT left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's just link MLIR PR to this PR for ref :))

@dgolubovicTT dgolubovicTT merged commit e405246 into main Dec 25, 2024
7 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants