Skip to content

Commit

Permalink
add new models in autotp workflow and enable gptj, baichuan
Browse files Browse the repository at this point in the history
  • Loading branch information
delock committed Feb 22, 2024
1 parent 4560381 commit e5b0b18
Showing 1 changed file with 90 additions and 0 deletions.
90 changes: 90 additions & 0 deletions .github/workflows/autotp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -126,3 +126,93 @@ jobs:
cd DeepSpeedExamples/inference/huggingface/text-generation
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
- name: AutoTP test (EleutherAI/gpt-j-6b)
run: |
export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
source oneCCL/build/_install/env/setvars.sh
# modify MODEL to change the model name, other lines are the same
export MODEL=EleutherAI/gpt-j-6b
cd DeepSpeedExamples/inference/huggingface/text-generation
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
- name: AutoTP test (baichuan-inc/Baichuan-7B)
run: |
export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
source oneCCL/build/_install/env/setvars.sh
# modify MODEL to change the model name, other lines are the same
export MODEL=baichuan-inc/Baichuan-7B
cd DeepSpeedExamples/inference/huggingface/text-generation
deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
# - name: AutoTP test (bigcode/starcoder)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=bigcode/starcoder
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (tiiuae/falcon-7b)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=tiiuae/falcon-7b
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (google/flan-t5-xl)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=google/flan-t5-xl
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (mistralai/Mistral-7B-v0.1)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=mistralai/Mistral-7B-v0.1
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (mosaicml/mpt-7b)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=mosaicml/mpt-7b
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (meta-llama/Llama-2-7b-hf)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=meta-llama/Llama-2-7b-hf
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1
#
# - name: AutoTP test (EleutherAI/gpt-neox-20b)
# run: |
# export LD_PRELOAD=/usr/lib/x86_64-linux-gnu/libstdc++.so.6
# source oneCCL/build/_install/env/setvars.sh
# # modify MODEL to change the model name, other lines are the same
# export MODEL=EleutherAI/gpt-neox-20
# cd DeepSpeedExamples/inference/huggingface/text-generation
# deepspeed --num_gpus 2 --bind_cores_to_rank inference-test.py --model $MODEL --dtype bfloat16 --use_meta_tensor
# deepspeed --num_gpus 2 --bind_cores_to_rank ds-hf-compare.py --model $MODEL --dtype bfloat16 --num_inputs 1

0 comments on commit e5b0b18

Please sign in to comment.