Skip to content

Commit

Permalink
undo mcli yaml change
Browse files Browse the repository at this point in the history
  • Loading branch information
vchiley committed Nov 10, 2023
1 parent a829f2c commit 1160b04
Showing 1 changed file with 7 additions and 13 deletions.
20 changes: 7 additions & 13 deletions mcli/mcli-1b-max-seq-len-8k.yaml
Original file line number Diff line number Diff line change
@@ -1,13 +1,10 @@
integrations:
- integration_type: git_repo
git_repo: vchiley/llm-foundry
git_branch: notie_embd
git_repo: mosaicml/llm-foundry
git_branch: v0.3.0
# git_commit: # OR use your commit hash
pip_install: -e .[gpu]
ssh_clone: false # Should be true if using a private repo
- integration_type: wandb
entity: mosaic-ml
project: notie_embd_test

# We are fetching, converting, and training on the 'val' split
# as it is small and quick to get going for this demo.
Expand All @@ -21,12 +18,10 @@ command: |
--concat_tokens 8192 --tokenizer EleutherAI/gpt-neox-20b --eos_text '<|endoftext|>'
composer train/train.py /mnt/config/parameters.yaml
image: mosaicml/pytorch:1.13.1_cu117-python3.10-ubuntu20.04

name: mpt-1b-ctx-8k-gpus-8-notieembd
name: mpt-1b-ctx-8k-gpus-8

compute:
gpus: 8 # Number of GPUs to use
cluster: r1z1

## These configurations are optional
# cluster: TODO # Name of the cluster to use for this run
Expand All @@ -53,7 +48,6 @@ parameters:
expansion_ratio: 4
max_seq_len: ${max_seq_len}
vocab_size: 50368
tie_word_embeddings: false
attn_config:
attn_impl: triton

Expand Down Expand Up @@ -108,7 +102,7 @@ parameters:
clipping_type: norm
clipping_threshold: 1.0

max_duration: 500ba # ~ 26B tokens
max_duration: 24800ba # ~ 26B tokens
eval_interval: 2000ba
eval_first: false
eval_subset_num_batches: -1
Expand All @@ -117,7 +111,7 @@ parameters:
# System
seed: 17
device_eval_batch_size: 1
device_train_microbatch_size: 4
device_train_microbatch_size: 1
# device_train_microbatch_size: auto
precision: amp_bf16

Expand All @@ -142,8 +136,8 @@ parameters:
lr_monitor: {}
memory_monitor: {}
runtime_estimator: {}
loggers:
wandb: {}
# loggers:
# wandb: {}

# Checkpoint to local filesystem or remote object store
# save_interval: 2000ba
Expand Down

0 comments on commit 1160b04

Please sign in to comment.