Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update_pretrain_benchmarks #543

Merged
merged 43 commits into from
Oct 19, 2023
Merged

Update_pretrain_benchmarks #543

merged 43 commits into from
Oct 19, 2023

Conversation

crinard
Copy link
Contributor

@crinard crinard commented Aug 22, 2023

Updates the benchmarking script for updates to LLM Foundry, Torch 2.0, Torch.compile, greater FSDP sweeps, allow H100 profiling and FP8, and add tables for H100 w/ BF16 & (very sparse) FP8.

@crinard crinard changed the title DRAFT: Update_pretrain_benchmarks Update_pretrain_benchmarks Oct 3, 2023
@vchiley vchiley requested review from hanlint and abhi-mosaic October 5, 2023 23:11
@vchiley
Copy link
Contributor

vchiley commented Oct 10, 2023

Comments about the table

Can we

  • separate out the table into 2 tables. one for amp_fp16 and another for amp_fp8
  • include a Model TFLOP column
    • MFU between different hardware is hard to compare
  • amp_fp8 is only used with MP Mode: DEFAULT; why not use PURE?
  • remove 350M results

scripts/train/benchmarking/README.md Show resolved Hide resolved
scripts/train/benchmarking/collect_results.py Show resolved Hide resolved
scripts/train/benchmarking/submit_benchmarks.py Outdated Show resolved Hide resolved
scripts/train/benchmarking/submit_benchmarks.py Outdated Show resolved Hide resolved
scripts/train/benchmarking/submit_benchmarks.py Outdated Show resolved Hide resolved
scripts/train/benchmarking/submit_benchmarks.py Outdated Show resolved Hide resolved
@vchiley vchiley merged commit b2a43a1 into mosaicml:main Oct 19, 2023
12 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants