Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Vram fix attempt #1164

Merged
merged 6 commits into from
Jan 23, 2024
Merged

Vram fix attempt #1164

merged 6 commits into from
Jan 23, 2024

Conversation

winglian
Copy link
Collaborator

@winglian winglian commented Jan 22, 2024

Description

more permanently fixes issues from #1127 dealing with VRAM instability. reverts an ordering change stemming from #1066 that was causing VRAM spikes during training. Updated dataset len calculations to use numpy instead of pandas since pandas can use the GPU and added a fallback calculation so we can still call it before dropping rows longer than the max length.

reverted order of calling process_datasets_for_packing from #1141 which can cause issues when changing max sequence length and needing to re-preprocess the datasets.

Motivation and Context

How has this been tested?

Screenshots (if appropriate)

Types of changes

Social Handles (Optional)

@winglian winglian merged commit 32580c1 into main Jan 23, 2024
6 of 7 checks passed
@winglian winglian deleted the vram-fix-attempt branch January 23, 2024 00:54
djsaunde pushed a commit that referenced this pull request Dec 17, 2024
* revert order of filter/drop_long step and handle calc for max_input_len only during preprocessing

* revert some changes to preparing for packing to allow more flexibility

* prepare dataset for packing during pre-processing step

* prepare dataset hash based on sample packing too

* enclose none check

* just cast straight to string for ds hash
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant