Fixing the gen_attention_mask_in_length function to handle the case when sequence id is -1 due to attention masking#940
Merged
ShashankMosaicML merged 32 commits intomosaicml:mainfrom ShashankMosaicML:shashank/fix_gen_attention_mask_in_length_attention_maskingFeb 5, 2024
+9
Commits
Commits on Oct 9, 2023
Commits on Oct 27, 2023
Commits on Nov 6, 2023
Commits on Nov 8, 2023
Commits on Nov 14, 2023
Commits on Nov 15, 2023
Commits on Dec 2, 2023
Commits on Dec 8, 2023
- committed
Commits on Dec 11, 2023
Commits on Dec 13, 2023
Commits on Dec 15, 2023
- committed
- committed
Commits on Dec 16, 2023
Commits on Dec 19, 2023
Commits on Dec 20, 2023
- committed
Commits on Jan 1, 2024
Commits on Jan 2, 2024
Commits on Jan 3, 2024
Commits on Jan 5, 2024
Commits on Jan 17, 2024
Commits on Jan 18, 2024
Commits on Jan 29, 2024
Commits on Feb 1, 2024
Commits on Feb 3, 2024
Commits on Feb 5, 2024
- committed
- authored
- committed
- authored