Skip to content

Fixing the gen_attention_mask_in_length function to handle the case when sequence id is -1 due to attention masking#940

Merged
ShashankMosaicML merged 32 commits intomosaicml:mainfrom ShashankMosaicML:shashank/fix_gen_attention_mask_in_length_attention_maskingFeb 5, 2024

Commits

Commits on Oct 9, 2023

Commits on Oct 27, 2023

Commits on Nov 6, 2023

Commits on Nov 8, 2023

Commits on Nov 14, 2023

Commits on Nov 15, 2023

Commits on Dec 2, 2023

Commits on Dec 8, 2023

Commits on Dec 11, 2023

Commits on Dec 13, 2023

Commits on Dec 15, 2023

Commits on Dec 16, 2023

Commits on Dec 19, 2023

Commits on Dec 20, 2023

Commits on Jan 1, 2024

Commits on Jan 2, 2024

Commits on Jan 3, 2024

Commits on Jan 5, 2024

Commits on Jan 17, 2024

Commits on Jan 18, 2024

Commits on Jan 29, 2024

Commits on Feb 1, 2024

Commits on Feb 3, 2024