Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[moe] merge branch feature/pipeline #4435

Merged
merged 88 commits into from
Aug 15, 2023
Merged

Commits on Jun 20, 2023

  1. [cluster] add process group mesh (hpcaitech#4039)

    * [cluster] add process group mesh
    
    * [test] add process group mesh test
    
    * force sync
    ver217 authored Jun 20, 2023
    Configuration menu
    Copy the full SHA
    1015f04 View commit details
    Browse the repository at this point in the history

Commits on Jun 27, 2023

  1. [pipeline] add stage manager (hpcaitech#4093)

    * [pipeline] add stage manager
    
    * [test] add pipeline stage manager test
    
    * [pipeline] add docstring for stage manager
    ver217 authored Jun 27, 2023
    Configuration menu
    Copy the full SHA
    b10821a View commit details
    Browse the repository at this point in the history

Commits on Jun 28, 2023

  1. [pipeline] implement p2p communication (hpcaitech#4100)

    * [pipeline] add p2p communication
    
    * [test] add p2p communication test
    
    * [test] add rerun decorator
    
    * [test] rename to avoid conflict
    ver217 authored Jun 28, 2023
    Configuration menu
    Copy the full SHA
    bd6b0a3 View commit details
    Browse the repository at this point in the history

Commits on Jun 29, 2023

  1. [pipeline] refactor 1f1b schedule (hpcaitech#4115)

    * [api] update optimizer wrapper to fit pipeline
    
    * [pipeline] add base schedule
    
    * [pipeline] add 1f1b schedule
    
    * [test] add pipeline schedule utils test
    
    * [pipeline] fix import
    ver217 authored Jun 29, 2023
    Configuration menu
    Copy the full SHA
    faeac9d View commit details
    Browse the repository at this point in the history

Commits on Jul 4, 2023

  1. [pipeline]add pipeline policy and bert forward (hpcaitech#4130)

    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    CjhHa1 authored Jul 4, 2023
    Configuration menu
    Copy the full SHA
    9852743 View commit details
    Browse the repository at this point in the history
  2. [cluster] add process group mesh (hpcaitech#4039)

    * [cluster] add process group mesh
    
    * [test] add process group mesh test
    
    * force sync
    ver217 committed Jul 4, 2023
    Configuration menu
    Copy the full SHA
    3be0c35 View commit details
    Browse the repository at this point in the history
  3. [pipeline] add stage manager (hpcaitech#4093)

    * [pipeline] add stage manager
    
    * [test] add pipeline stage manager test
    
    * [pipeline] add docstring for stage manager
    ver217 committed Jul 4, 2023
    Configuration menu
    Copy the full SHA
    18c7539 View commit details
    Browse the repository at this point in the history
  4. [pipeline] implement p2p communication (hpcaitech#4100)

    * [pipeline] add p2p communication
    
    * [test] add p2p communication test
    
    * [test] add rerun decorator
    
    * [test] rename to avoid conflict
    ver217 committed Jul 4, 2023
    Configuration menu
    Copy the full SHA
    5a467e9 View commit details
    Browse the repository at this point in the history
  5. [pipeline] refactor 1f1b schedule (hpcaitech#4115)

    * [api] update optimizer wrapper to fit pipeline
    
    * [pipeline] add base schedule
    
    * [pipeline] add 1f1b schedule
    
    * [test] add pipeline schedule utils test
    
    * [pipeline] fix import
    ver217 committed Jul 4, 2023
    Configuration menu
    Copy the full SHA
    9526f44 View commit details
    Browse the repository at this point in the history
  6. [pipeline]add pipeline policy and bert forward (hpcaitech#4130)

    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    CjhHa1 authored and ver217 committed Jul 4, 2023
    Configuration menu
    Copy the full SHA
    836a3a2 View commit details
    Browse the repository at this point in the history
  7. Merge pull request hpcaitech#4166 from ver217/sync/main

    [sync] update from main
    FrankLeeeee authored Jul 4, 2023
    Configuration menu
    Copy the full SHA
    ef1f972 View commit details
    Browse the repository at this point in the history

Commits on Jul 5, 2023

  1. [pipeline] build bloom model and policy , revise the base class of po…

    …licy (hpcaitech#4161)
    
    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    
    * add bloom model and policy ,revise the base class of policy
    
    * revise
    
    * revision
    
    * add bert_for_pretraining
    CjhHa1 authored Jul 5, 2023
    Configuration menu
    Copy the full SHA
    386d34e View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    15a4e82 View commit details
    Browse the repository at this point in the history
  3. Configuration menu
    Copy the full SHA
    8b6679d View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    9143556 View commit details
    Browse the repository at this point in the history
  5. [test] add shard util tests

    ver217 committed Jul 5, 2023
    Configuration menu
    Copy the full SHA
    0cbe423 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    1a87dd7 View commit details
    Browse the repository at this point in the history
  7. [shardformer] fix type hint

    ver217 committed Jul 5, 2023
    Configuration menu
    Copy the full SHA
    d4b96ab View commit details
    Browse the repository at this point in the history
  8. Merge pull request hpcaitech#4176 from ver217/feature/pipeline-policy

    [pipeline] fit shardformer policy
    CjhHa1 authored Jul 5, 2023
    Configuration menu
    Copy the full SHA
    12e6d5d View commit details
    Browse the repository at this point in the history

Commits on Jul 6, 2023

  1. [pipeline] add bert_for_pretraining bert_lmhead forward and policy (h…

    …pcaitech#4172)
    
    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    
    * add bloom model and policy ,revise the base class of policy
    
    * revise
    
    * revision
    
    * add bert_for_pretraining
    
    * add bert_for_pretraining forward and policy
    
    * fix typos
    
    * cancel warning
    
    * change the imediate output to default dict
    
    * change the default output of get_shared_params
    CjhHa1 authored Jul 6, 2023
    Configuration menu
    Copy the full SHA
    15b34e0 View commit details
    Browse the repository at this point in the history

Commits on Jul 7, 2023

  1. Feature/vit support (hpcaitech#4182)

    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    klhhhhh authored Jul 7, 2023
    Configuration menu
    Copy the full SHA
    ec217de View commit details
    Browse the repository at this point in the history
  2. [pipeline] move bert related pipeline components to shardformer (hpca…

    …itech#4187)
    
    * move bert related pipeline components to shardformer
    
    * fix bugs
    
    * revision
    
    * fix bert model tests
    
    * fix bert_lm_head model tests
    
    * fix tests
    
    * fix tests
    
    * done checks
    
    * skip bloom
    CjhHa1 authored Jul 7, 2023
    Configuration menu
    Copy the full SHA
    c6f9c2c View commit details
    Browse the repository at this point in the history

Commits on Jul 10, 2023

  1. [shardformer] support lazy init (hpcaitech#4202)

    * [shardformer] support lazy init
    
    * [shardformer] linear support lazy init
    
    * [shardformer] embedding support lazy init
    
    * [shardformer] norm support lazy init
    
    * [shardformer] fused linear support lazy init
    
    * [test] update shardformer test layer
    
    * [test] shardformer with lazy init fit ddp
    
    * [lazy] hotfix deepcopy of param
    
    * [shardformer] fix bert policy and update test
    
    * [shardformer] fix bloom policy and update test
    
    * [shardformer] fix opt policy and update test
    
    * [shardformer] fix t5 policy and update test
    
    * [shardformer] fix gpt2 policy and update test
    
    * [shardformer] fix llama policy and update test
    ver217 authored Jul 10, 2023
    Configuration menu
    Copy the full SHA
    0192011 View commit details
    Browse the repository at this point in the history
  2. [pipeline] Bert pipeline for shardformer and its tests (hpcaitech#4197)

    * add pipeline forward
    
    * complete pipeline forward check
    
    * fix bert forward without pipeline
    
    * fix comments
    
    * discard useless line
    
    * add todo
    
    * clean prints
    
    * fix distribute layers
    CjhHa1 authored Jul 10, 2023
    Configuration menu
    Copy the full SHA
    b30d1b9 View commit details
    Browse the repository at this point in the history

Commits on Jul 11, 2023

  1. [pipeline] Llama pipeline (hpcaitech#4205)

    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * Revert "bloom policy"
    
    This reverts commit 8dee68a.
    
    This policy should be revert and copied to feature/bloom
    
    * revert the bloom changes
    
    * cancel unneeded inputs
    
    * gpt
    CjhHa1 authored Jul 11, 2023
    Configuration menu
    Copy the full SHA
    a2619c3 View commit details
    Browse the repository at this point in the history
  2. [pipeline] Llama causal lm and llama for sequence classification pipe…

    …line (hpcaitech#4208)
    
    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * Revert "bloom policy"
    
    This reverts commit 8dee68a.
    
    This policy should be revert and copied to feature/bloom
    
    * revert the bloom changes
    
    * cancel unneeded inputs
    
    * gpt
    
    * finish llama
    
    * causal lm and sequence classification
    
    * revision
    CjhHa1 authored Jul 11, 2023
    Configuration menu
    Copy the full SHA
    981764c View commit details
    Browse the repository at this point in the history

Commits on Jul 13, 2023

  1. [pipeline] add bloom model pipeline (hpcaitech#4210)

    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * finish bloom model
    
    * test shard gpt2
    
    * clear cache
    CjhHa1 authored Jul 13, 2023
    Configuration menu
    Copy the full SHA
    3595eba View commit details
    Browse the repository at this point in the history
  2. [pipeline] Add Pipeline Forward for GPT2Model Shardformer (hpcaitech#…

    …4224)
    
    * * fix typehint & docstring in sharder.py
    
    * * update pipeline forward for GPT2Model
    
    * * add test for pipeline forward of GPT2Model
    
    * * add cache cleaning in gpt2 test
    
    * * change assert to raise command
    Fridge003 authored Jul 13, 2023
    Configuration menu
    Copy the full SHA
    236f294 View commit details
    Browse the repository at this point in the history

Commits on Jul 14, 2023

  1. Configuration menu
    Copy the full SHA
    ad2687c View commit details
    Browse the repository at this point in the history
  2. [shardformer] support SAM (hpcaitech#4231)

    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    FoolPlayer authored Jul 14, 2023
    Configuration menu
    Copy the full SHA
    ddecf73 View commit details
    Browse the repository at this point in the history

Commits on Jul 17, 2023

  1. [shardformer] support whisper (hpcaitech#4212)

    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    FoolPlayer authored Jul 17, 2023
    Configuration menu
    Copy the full SHA
    afcf4a0 View commit details
    Browse the repository at this point in the history
  2. [pipeline] add pipeline forward for variants of gpt2 (hpcaitech#4238)

    * add forward for GPTLMHeadModel
    
    * add test for gpt_lm
    
    * arranging get_held_layers method
    
    * arrange forward replacement
    
    * add forward for GPT2ForTokenClassification
    
    * add forward for GPT2ForSequenceClassification
    
    * fix test_shard_gpt2.py
    
    * add GPT2DoubleHeadsmodel & fix bugs
    
    * add id checking in get_shared_params
    Fridge003 authored Jul 17, 2023
    Configuration menu
    Copy the full SHA
    383d2e3 View commit details
    Browse the repository at this point in the history
  3. [pipeline] All bert models (hpcaitech#4233)

    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * Revert "bloom policy"
    
    This reverts commit 8dee68a.
    
    This policy should be revert and copied to feature/bloom
    
    * revert the bloom changes
    
    * cancel unneeded inputs
    
    * gpt
    
    * finish llama
    
    * causal lm and sequence classification
    
    * revision
    
    * add pure pipeline test
    
    * finish some bert models
    
    * finish all bert models
    
    * finish bert tests
    
    * fix bugs
    
    * fix bugs
    
    * fix test pipeline
    
    * fix data gen for qa
    
    * update the set pipeline forward
    
    * shared params
    
    * fix bugs
    CjhHa1 authored Jul 17, 2023
    Configuration menu
    Copy the full SHA
    7b8756f View commit details
    Browse the repository at this point in the history
  4. [pipeline] finish bloom models pipeline and tests (hpcaitech#4223)

    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * finish bloom model
    
    * test shard gpt2
    
    * clear cache
    
    * support all bloom models
    
    * add bloom models policies
    
    * finish bloom pipeline and tests
    
    * add set pipeline
    
    * finish bloom
    CjhHa1 authored Jul 17, 2023
    Configuration menu
    Copy the full SHA
    a895458 View commit details
    Browse the repository at this point in the history

Commits on Jul 18, 2023

  1. [bugs] hot fix some testing bugs for new models (hpcaitech#4268)

    * hot fix
    
    * hot fx tracer
    CjhHa1 authored Jul 18, 2023
    Configuration menu
    Copy the full SHA
    843158b View commit details
    Browse the repository at this point in the history

Commits on Jul 19, 2023

  1. [pipeline] support shardformer for GPT2ForQuestionAnswering & complet…

    …e pipeline support for GPT2 (hpcaitech#4245)
    
    * change for transformers loggers
    
    * add forward for GPT2ForQuestionAnswering
    
    * fix assert
    
    * fix torchrec test
    Fridge003 authored Jul 19, 2023
    Configuration menu
    Copy the full SHA
    3918898 View commit details
    Browse the repository at this point in the history

Commits on Jul 20, 2023

  1. [shardformer] support inplace sharding (hpcaitech#4251)

    * [shardformer] embedding support inplace sharding
    
    * [shardformer] linear support inplace sharding
    
    * [shardformer] layernorm support inplace sharding
    
    * [shardformer] qkv support inplace sharding
    
    * [test] update shardformer layer test
    
    * [shardformer] fix shared param sharding
    
    * [shardformer] fix bert policy
    
    * [shardformer] fix bloom policy
    
    * [shardformer] fix llama policy
    
    * [shardformer] fix opt policy
    
    * [shardformer] fix t5 policy
    
    * [shardformer] fix fused qkv linear
    
    * [shardformer] fix bugs
    
    * force sync
    
    * [test] fix bugs
    
    * [test] fix transformer version
    ver217 authored Jul 20, 2023
    Configuration menu
    Copy the full SHA
    7b5a155 View commit details
    Browse the repository at this point in the history
  2. [pipeline] refactor gpt2 pipeline forwards (hpcaitech#4287)

    * move gpt2 pipeline forwards to modeling folder
    
    * check pipeline status when adding replacing policy
    
    * fix typehint
    
    * fix arguments processing in gpt2_model_forward
    Fridge003 authored Jul 20, 2023
    Configuration menu
    Copy the full SHA
    cc120c6 View commit details
    Browse the repository at this point in the history
  3. [pipeline] OPT model pipeline (hpcaitech#4258)

    * opt forward and test
    
    * pause
    
    * finish opt model pipeline
    
    * finish opt pipeline
    
    * opt forward and test
    
    * pause
    
    * finish opt model pipeline
    
    * finish opt pipeline
    
    * fix opt
    
    * set transformers version
    
    * refactor the test pipeline
    CjhHa1 authored Jul 20, 2023
    Configuration menu
    Copy the full SHA
    7b583c0 View commit details
    Browse the repository at this point in the history
  4. [hotfix] fix opt pipeline (hpcaitech#4293)

    * opt forward and test
    
    * pause
    
    * finish opt model pipeline
    
    * finish opt pipeline
    
    * opt forward and test
    
    * pause
    
    * finish opt model pipeline
    
    * finish opt pipeline
    
    * fix opt
    
    * set transformers version
    
    * refactor the test pipeline
    
    * fix bug
    CjhHa1 authored Jul 20, 2023
    Configuration menu
    Copy the full SHA
    3b92e4a View commit details
    Browse the repository at this point in the history
  5. Feature/chatglm (hpcaitech#4240)

    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    klhhhhh authored Jul 20, 2023
    Configuration menu
    Copy the full SHA
    77cc087 View commit details
    Browse the repository at this point in the history
  6. [shardformer] added tests

    klhhhhh committed Jul 20, 2023
    Configuration menu
    Copy the full SHA
    6c2acf0 View commit details
    Browse the repository at this point in the history
  7. Configuration menu
    Copy the full SHA
    7668b24 View commit details
    Browse the repository at this point in the history
  8. import chatglm

    klhhhhh committed Jul 20, 2023
    Configuration menu
    Copy the full SHA
    b135b75 View commit details
    Browse the repository at this point in the history
  9. Configuration menu
    Copy the full SHA
    e3cd5cb View commit details
    Browse the repository at this point in the history
  10. Configuration menu
    Copy the full SHA
    30574a7 View commit details
    Browse the repository at this point in the history
  11. Configuration menu
    Copy the full SHA
    28677d4 View commit details
    Browse the repository at this point in the history
  12. [shardformer] polish code

    klhhhhh committed Jul 20, 2023
    Configuration menu
    Copy the full SHA
    28319c2 View commit details
    Browse the repository at this point in the history
  13. Configuration menu
    Copy the full SHA
    3f19de9 View commit details
    Browse the repository at this point in the history
  14. Configuration menu
    Copy the full SHA
    2a4bbcf View commit details
    Browse the repository at this point in the history
  15. Configuration menu
    Copy the full SHA
    32448e3 View commit details
    Browse the repository at this point in the history
  16. Configuration menu
    Copy the full SHA
    eb1c71a View commit details
    Browse the repository at this point in the history
  17. Configuration menu
    Copy the full SHA
    127e385 View commit details
    Browse the repository at this point in the history
  18. Configuration menu
    Copy the full SHA
    9d5b141 View commit details
    Browse the repository at this point in the history

Commits on Jul 21, 2023

  1. [pipeline] reformat for unified design (hpcaitech#4283)

    * bert_reformat
    
    * reformat
    
    * reformat
    
    * fix a typo
    
    * format
    
    * format
    
    * fix bug
    CjhHa1 authored Jul 21, 2023
    Configuration menu
    Copy the full SHA
    d7e584c View commit details
    Browse the repository at this point in the history
  2. [pipeline] add pipeline support for T5Stack/T5EncoderModel (hpcaitech…

    …#4300)
    
    * modify t5 policy & add test
    
    * pipeline stage distribution for t5
    
    * complete t5 base policy
    
    * t5 stack: halfway
    
    * modify gpt2 pipeline test
    
    * complete pipeline forward for T5Stack/T5EncoderModel
    
    * fix docstring
    
    * move t5 util tests to test_pipeline
    Fridge003 authored Jul 21, 2023
    Configuration menu
    Copy the full SHA
    9605805 View commit details
    Browse the repository at this point in the history
  3. Merge pull request hpcaitech#4297 from klhhhhh/feature/support_ChatGL…

    …MForConditionalGeneration
    
    Feature/support chat glm for conditional generation
    klhhhhh authored Jul 21, 2023
    Configuration menu
    Copy the full SHA
    805f342 View commit details
    Browse the repository at this point in the history

Commits on Jul 25, 2023

  1. [shardformer] support Blip2 (hpcaitech#4243)

    * support base blip2
    
    * add support for downstream blip2 model
    
    * update readme
    
    * add forward injection
    
    * skip not compatible models test
    
    * fix test for gemini and low_level_zero_pugin
    FoolPlayer authored Jul 25, 2023
    Configuration menu
    Copy the full SHA
    f48a8bb View commit details
    Browse the repository at this point in the history
  2. [pipeline] test pure pipeline process using llama (hpcaitech#4218)

    * bloom policy
    
    * llama pipeline forward and tests
    
    * fix the output and attention_mask
    
    * fix name
    
    * bind argument to policy
    
    * Revert "bloom policy"
    
    This reverts commit 8dee68a.
    
    This policy should be revert and copied to feature/bloom
    
    * revert the bloom changes
    
    * cancel unneeded inputs
    
    * gpt
    
    * finish llama
    
    * causal lm and sequence classification
    
    * revision
    
    * add pure pipeline test
    
    * fixed version
    
    * fixed version
    
    * pure pipeline
    CjhHa1 authored Jul 25, 2023
    Configuration menu
    Copy the full SHA
    965bf20 View commit details
    Browse the repository at this point in the history
  3. [pipeline] add pipeline support for all T5 models (hpcaitech#4310)

    * complete policy for T5Model & T5ForConditionalGeneration
    
    * modify function signature in forwards
    
    * add forward for T5model
    
    * add forward for T5ForConditionalGeneration
    
    * fix a bug
    
    * fix hidden_states transporting in decoder
    
    * fix the passing of encoder_outputs
    Fridge003 authored Jul 25, 2023
    Configuration menu
    Copy the full SHA
    28e6980 View commit details
    Browse the repository at this point in the history
  4. [shardformer] support pipeline base vit model (hpcaitech#4284)

    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * support base vit pipeline
    
    * support vit downstream model
    
    * fix vit shard test
    
    * modify hidden states return type
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    FoolPlayer and klhhhhh authored Jul 25, 2023
    Configuration menu
    Copy the full SHA
    2e93d9b View commit details
    Browse the repository at this point in the history
  5. [plugin] add 3d parallel plugin (hpcaitech#4295)

    * [amp] add mixed precision optimizer
    
    * [plugin] add 3d parallel plugin
    
    * [booster] support pipeline
    
    * [plugin] 3d parallel plugin support clip grad norm
    
    * [shardformer] fix sharder and add plugin test
    
    * [plugin] rename 3d parallel plugin
    
    * [ci] support testmon core pkg change detection (hpcaitech#4305)
    
    * [hotfix] debug testmon
    
    * [hotfix] fix llama
    
    * [hotfix] fix p2p bugs
    
    * [hotfix] fix requirements
    ver217 authored Jul 25, 2023
    Configuration menu
    Copy the full SHA
    78dd508 View commit details
    Browse the repository at this point in the history

Commits on Jul 27, 2023

  1. [hotfix] fix gemini and zero test (hpcaitech#4333)

    * [hotfix] fix gemini and zero test
    
    * [hotfix] fix lazy init test
    
    * [hotfix] fix lazy init test
    ver217 authored Jul 27, 2023
    Configuration menu
    Copy the full SHA
    8ad05d1 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    d547377 View commit details
    Browse the repository at this point in the history

Commits on Jul 31, 2023

  1. [pipeline] add unit test for 1f1b (hpcaitech#4303)

    * add unit test for 1f1b
    
    * polish code
    
    * polish code and update ut version
    
    * fix
    Gy-Lu authored Jul 31, 2023
    Configuration menu
    Copy the full SHA
    b941e65 View commit details
    Browse the repository at this point in the history

Commits on Aug 1, 2023

  1. [pipeline] refactor test pipeline and remove useless utils in pipeline (

    hpcaitech#4324)
    
    * refactor tests
    
    * refactor bloom model
    
    * finish policy tests
    
    * refactor tests
    
    * fix test pure pipeline
    
    * remove test pipeline and cutdown launch process
    
    * refactor tests
    
    * refactor bloom model
    
    * finish policy tests
    
    * refactor tests
    
    * fix test pure pipeline
    
    * remove test pipeline and cutdown launch process
    CjhHa1 authored Aug 1, 2023
    Configuration menu
    Copy the full SHA
    7d5b144 View commit details
    Browse the repository at this point in the history
  2. Configuration menu
    Copy the full SHA
    01ef6c5 View commit details
    Browse the repository at this point in the history
  3. [pipeline] support fp32 for HybridPlugin/merge shardformer test and p…

    …ipeline test into one file (hpcaitech#4354)
    
    * add naive optimizer for 3DPlugin/refactor gpt2 shardformer test
    
    * merge tests of PP/DP/TP combinations into one test file
    
    * fix bug when sync grad for dp in HybridPlugin
    
    * update supported precisions for 3DPlugin/fix bug when shifting tp_degree
    
    * improve the passing of lazy_init
    
    * modify lazy_init/use sync_shared_params
    Fridge003 authored Aug 1, 2023
    Configuration menu
    Copy the full SHA
    992cbb7 View commit details
    Browse the repository at this point in the history
  4. Configuration menu
    Copy the full SHA
    260df9e View commit details
    Browse the repository at this point in the history

Commits on Aug 2, 2023

  1. Configuration menu
    Copy the full SHA
    5403578 View commit details
    Browse the repository at this point in the history
  2. Merge pull request hpcaitech#4358 from hpcaitech/feature/shardformer-…

    …models
    
    [Merge] Feature/shardformer models to feature/pipeline
    FrankLeeeee authored Aug 2, 2023
    Configuration menu
    Copy the full SHA
    b849657 View commit details
    Browse the repository at this point in the history

Commits on Aug 3, 2023

  1. [test] Hotfix/fix some model test and refactor check util api (hpcait…

    …ech#4369)
    
    * fix llama test
    
    * fix test bug of bert, blip2, bloom, gpt2
    
    * fix llama test
    
    * fix opt test
    
    * fix sam test
    
    * fix sam test
    
    * fix t5 test
    
    * fix vit test
    
    * fix whisper test
    
    * fix whisper test
    
    * polish code
    
    * adjust allclose parameter
    
    * Add mistakenly deleted code
    
    * addjust allclose
    
    * change loss function for some base model
    FoolPlayer authored Aug 3, 2023
    Configuration menu
    Copy the full SHA
    3bfdd53 View commit details
    Browse the repository at this point in the history
  2. [shardformer] add util functions for shardformer tests/fix sync_share…

    …d_param (hpcaitech#4366)
    
    * add util functions for shardformer tests & rewrite gpt2 test
    
    * fix shared_params & embedding/merging
    
    * fix precision
    Fridge003 authored Aug 3, 2023
    Configuration menu
    Copy the full SHA
    21c6bb0 View commit details
    Browse the repository at this point in the history

Commits on Aug 4, 2023

  1. [pipeline] add chatglm (hpcaitech#4363)

    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    
    * add bloom model and policy ,revise the base class of policy
    
    * revise
    
    * revision
    
    * add bert_for_pretraining
    
    * add bert_for_pretraining forward and policy
    
    * fix typos
    
    * cancel warning
    
    * change the imediate output to default dict
    
    * change the default output of get_shared_params
    
    * add chatglm
    
    * add
    
    * chatglm
    
    * chatglm
    
    * finish chatglm
    
    * deletes
    
    * fix rmsnorm
    
    * chatglm
    
    * fix chatglm shard
    
    * init
    CjhHa1 authored Aug 4, 2023
    Configuration menu
    Copy the full SHA
    c5f4844 View commit details
    Browse the repository at this point in the history

Commits on Aug 7, 2023

  1. [Shardformer] Merge flash attention branch to pipeline branch (hpcait…

    …ech#4362)
    
    * [shardformer] supported flash attention test dependency (hpcaitech#4158)
    
    * [shardformer] fix flash attention utils test (hpcaitech#4180)
    
    * [shardformer] opt support flash attention (hpcaitech#4163)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] add performance benchmark of shardformer (hpcaitech#4175)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] benchmark fix
    
    * [shardformer] benchmark fix
    
    * [shardformer] llama support flash attention (hpcaitech#4185)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] llama support flash attention
    
    * [shardformer] llama support flash attention
    
    * [shardformer] Move the import statement for xformer outside the forward function.
    
    * [shardformer] gpt2 support flash attention. (hpcaitech#4191)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] gpt2 support flash attention
    
    * [shardformer] gpt2 support flash attention
    
    * [shardformer] gpt2 support flash attention
    
    * [shardformer] bloom support flash attention (hpcaitech#4188)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] bloom suport flash attention
    
    * [shardformer] add assert to sequence length
    
    * [shardformer] fix
    
    * [shardformer] fix
    
    * [shardformer] fix
    
    * [shardformer] bert support flash attention. (hpcaitech#4206)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] bert support flash attention
    
    * [shardformer] t5 support flash attention. (hpcaitech#4216)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] t5 support flash attention
    
    * [shardformer] t5 support flash attention
    
    * fix typo
    
    * fix typo
    
    * fix typo
    
    * fix typo
    
    * fix typo
    
    * fix typo
    
    * [shardformer] support 'paddedcausal'  type of attention mask in Coloattention. (hpcaitech#4215)
    
    * added padded causal attn mask type for ColoAttention
    
    * [shardformer]t5 flash attention fix (hpcaitech#4239)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] t5 flash attention fix
    
    * [shardformer] update gpt2 to use coloattention. (hpcaitech#4234)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] update gpt2 to use coloattention
    
    * [shardformer] update gpt2 to use coloattention
    
    * [shardformer] update gpt2 to use coloattention
    
    * [shardformer] update gpt2 to use coloattention
    
    * [shardformer] update gpt2
    
    * [shardformer] update opt and llama to use coloattention. (hpcaitech#4226)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt to use coloattention
    
    * [shardformer]update opt
    
    * [shardformer] shardformer support jit fused operator. (hpcaitech#4236)
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] opt support flash attention
    
    * [shardformer] move to modeling
    
    * [shardformer] move to modeling
    
    * [shardformer] bloom support jit fused operator
    
    * [shardformer] bloom support jit fused operator
    
    * [shardformer] bloom support jit fused operator
    
    * [shardformer] t5 support jit fused operator
    
    * [shardformer] t5 support jit fused operator
    
    * [shardformer] t5 support jit fused operator
    
    * [shardformer] add roadmap of flash attention
    
    * [shardformer] add roadmap of flash attention
    
    * [shardformer] add roadmap of flash attention
    
    * [shardformer] add type hint to 'self' param of forward
    
    * [shardformer] merge feature/shardformer-models branch to feature/flash-attention-shardformer branch. (hpcaitech#4290)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    
    * [shardformer] whisper support flash attention (hpcaitech#4301)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] whisper support flash attention
    
    * [shardformer] whisper support flash attention
    
    * [shardformer]whisper support jit operator
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    
    * [shardformer] sam support flash attention (hpcaitech#4316)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] sam support flash attention
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    
    * [shardformer] merge blip2/chatglm  (hpcaitech#4321)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] support ChatGLMForConditionalGeneration & add fusedlayernorm for vit
    
    * [shardformer] support Blip2 (hpcaitech#4243)
    
    * support base blip2
    
    * add support for downstream blip2 model
    
    * update readme
    
    * add forward injection
    
    * skip not compatible models test
    
    * fix test for gemini and low_level_zero_pugin
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    Co-authored-by: klhhhhh <[email protected]>
    
    * [shardformer] blip2 support flash attention and jit operator (hpcaitech#4325)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] support ChatGLMForConditionalGeneration & add fusedlayernorm for vit
    
    * [shardformer] support Blip2 (hpcaitech#4243)
    
    * support base blip2
    
    * add support for downstream blip2 model
    
    * update readme
    
    * add forward injection
    
    * skip not compatible models test
    
    * fix test for gemini and low_level_zero_pugin
    
    * [shardformer] blip2 support flash attention and jit operator
    
    * [shardformer] blip2 support flash attention and jit operator
    
    * [shardformer] blip2 support flash attention and jit operator
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    Co-authored-by: klhhhhh <[email protected]>
    
    * [shardformer] chatglm support flash attention and jit operator (hpcaitech#4330)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] support ChatGLMForConditionalGeneration & add fusedlayernorm for vit
    
    * [shardformer] support Blip2 (hpcaitech#4243)
    
    * support base blip2
    
    * add support for downstream blip2 model
    
    * update readme
    
    * add forward injection
    
    * skip not compatible models test
    
    * fix test for gemini and low_level_zero_pugin
    
    * [shardformer] chatglm support flash attention and jit operator
    
    * [shardformer] chatglm support flash attention and jit operator
    
    * [shardformer] chatglm support flash attention and jit operator
    
    * [shardformer] chatglm support flash attention and jit operator
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    Co-authored-by: klhhhhh <[email protected]>
    
    * [shardformer] vit support flash attention and jit operator (hpcaitech#4334)
    
    * Feature/vit support (hpcaitech#4182)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * fix attention dropout
    
    * [shardformer] support SAM (hpcaitech#4231)
    
    * 1.support sam 2.add fused qkv for nn.Linear
    
    * update utils support set element in list
    
    * overtwrite SamVisionAttention foward to use DropoutForParallelInput
    
    * remove unused code
    
    * [shardformer] support whisper (hpcaitech#4212)
    
    * support whisper
    
    * fix bug in vocabembedding
    
    * support downstream model of whisper
    
    * update readme
    
    * Feature/chatglm (hpcaitech#4240)
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * [shardformer] chatglm ready
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] chatglm shard without mlp sharding
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] fix chatglm configuration with pre-commit
    
    * [shardformer] added tests
    
    * [shardformer] vit test finish and support
    
    * import chatglm
    
    * [shardformer] add test kit in model zoo for chatglm
    
    * [sharformer] add first version of policy of chatglm
    
    * [shardformer] polish chatglm code
    
    * [shardformer] polish code
    
    * [shardformer] support chatglm without layernorm
    
    * [shardformer] delete some file
    
    * [shardformer] ChatGLM support layernorm sharding
    
    * [shardformer] register without auto policy
    
    * [shardformer] pre-commit check files
    
    * [shardformer] support ChatGLMForConditionalGeneration & add fusedlayernorm for vit
    
    * [shardformer] support Blip2 (hpcaitech#4243)
    
    * support base blip2
    
    * add support for downstream blip2 model
    
    * update readme
    
    * add forward injection
    
    * skip not compatible models test
    
    * fix test for gemini and low_level_zero_pugin
    
    * [shardformer] vit support flash attention and jit operator
    
    * [shardformer] vit support flash attention and jit operator
    
    ---------
    
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    Co-authored-by: klhhhhh <[email protected]>
    
    * [pipeline] merge flash attention branch
    
    * [pipeline] merge flash attention branch
    
    * [pipeline] merge flash attention branch
    
    * [pipeline] fix conflict
    
    * [pipeline] fix conflict
    
    * Merge branch 'feature/pipeline' into feature/pipeline
    
    * Merge branch 'feature/pipeline' into feature/pipeline
    
    * Merge branch 'feature/pipeline' into feature/pipeline
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * activate checks
    
    * fix flash attention tests
    
    * gemini ignore whisper
    
    * fix vit
    
    * fix xformers import handle
    
    ---------
    
    Co-authored-by: Frank Lee <[email protected]>
    Co-authored-by: Kun Lin <[email protected]>
    Co-authored-by: FoolPlayer <[email protected]>
    Co-authored-by: klhhhhh <[email protected]>
    5 people authored Aug 7, 2023
    Configuration menu
    Copy the full SHA
    7c84f51 View commit details
    Browse the repository at this point in the history

Commits on Aug 8, 2023

  1. [pipeline] rewrite t5 tests & support multi-tensor transmitting in pi…

    …peline (hpcaitech#4388)
    
    * fix remaining t5 bugs/rewrite t5 tests
    
    * fix multi-tensor communication in pipeline
    
    * rearrange test_config
    
    * fix keyerror in sync_shared_params
    
    * fix get_held_layers & Randomnizer, complete t5 tests
    
    * erase printing
    
    * fix get_held_layers through modifying _release_unheld_layers
    
    * fix _get_recursive_held_layers bug
    Fridge003 authored Aug 8, 2023
    Configuration menu
    Copy the full SHA
    2e77e57 View commit details
    Browse the repository at this point in the history

Commits on Aug 9, 2023

  1. [shardformer] update shardformer to use flash attention 2 (hpcaitech#…

    …4392)
    
    * cherry-pick flash attention 2
    
    cherry-pick flash attention 2
    
    * [shardformer] update shardformer to use flash attention 2
    
    [shardformer] update shardformer to use flash attention 2, fix
    
    [shardformer] update shardformer to use flash attention 2, fix
    
    [shardformer] update shardformer to use flash attention 2, fix
    flybird11111 authored Aug 9, 2023
    Configuration menu
    Copy the full SHA
    c14920a View commit details
    Browse the repository at this point in the history

Commits on Aug 10, 2023

  1. [shardformer] test all optimizations (hpcaitech#4399)

    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    flybird11111 authored Aug 10, 2023
    Configuration menu
    Copy the full SHA
    ed2c229 View commit details
    Browse the repository at this point in the history

Commits on Aug 11, 2023

  1. [pipeline] rewrite bert tests and fix some bugs (hpcaitech#4409)

    * add pipeline policy and bert forward to be done
    
    * add bertmodel pipeline forward and make tests
    
    * add Bert_Policy and test for policy
    
    * update formatting
    
    * update formatting
    
    * update the code
    
    * fix bugs
    
    * fix name confilt
    
    * add bloom model and policy ,revise the base class of policy
    
    * revise
    
    * revision
    
    * add bert_for_pretraining
    
    * add bert_for_pretraining forward and policy
    
    * fix typos
    
    * cancel warning
    
    * change the imediate output to default dict
    
    * change the default output of get_shared_params
    
    * rewrite bert test
    
    * rewrite bert test
    
    * fix some bugs
    
    * del pipeline tests
    
    * del pipeline tests
    
    * del useless print
    
    * del useless print
    
    * rewrite data repeats
    CjhHa1 authored Aug 11, 2023
    Configuration menu
    Copy the full SHA
    9916a19 View commit details
    Browse the repository at this point in the history
  2. [shardformer]fix, test gpt2 for AMP+TP (hpcaitech#4403)

    * [shardformer] gpt2 tests fix
    
    [shardformer] test all optimizations (hpcaitech#4399)
    
    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    
    [shardformer] gpt2 tests fix
    
    * [shardformer] gpt2 tests fix
    flybird11111 authored Aug 11, 2023
    Configuration menu
    Copy the full SHA
    fcbf80f View commit details
    Browse the repository at this point in the history
  3. [shardformer] rewrite tests for opt/bloom/llama/vit/chatglm (hpcaitec…

    …h#4395)
    
    * rewrite opt tests
    
    * rewrite llama tests
    
    * rewrite bloom & vit tests
    
    * rewrite chatglm tests
    
    * fix LinearCol for classfiers
    
    * add judge for other tp layers, fix lazy init in util
    Fridge003 authored Aug 11, 2023
    Configuration menu
    Copy the full SHA
    1e518ae View commit details
    Browse the repository at this point in the history
  4. [shardformer] update tests for all optimization (hpcaitech#4413)

    [shardformer] update tests for all optimization
    flybird11111 authored Aug 11, 2023
    Configuration menu
    Copy the full SHA
    d4a3a10 View commit details
    Browse the repository at this point in the history

Commits on Aug 14, 2023

  1. Configuration menu
    Copy the full SHA
    6990477 View commit details
    Browse the repository at this point in the history
  2. [shardformer]update t5 tests for using all optimizations. (hpcaitech#…

    …4407)
    
    * [shardformer] gpt2 tests fix
    
    [shardformer] test all optimizations (hpcaitech#4399)
    
    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    
    [shardformer] test all optimizations
    
    [shardformer] gpt2 tests fix
    
    * [shardformer]update t5 to use all optimizations
    flybird11111 authored Aug 14, 2023
    Configuration menu
    Copy the full SHA
    ac8d4ed View commit details
    Browse the repository at this point in the history
  3. [shardformer] update bloom/llama/vit/chatglm tests (hpcaitech#4420)

    [shardformer] update bloom/llama/vit/chatglm tests
    
    [shardformer] update opt tests
    
    [shardformer] update opt tests
    
    [shardformer] update bloom/llama/vit/chatglm tests
    
    [shardformer] update bloom/llama/vit/chatglm tests
    
    [shardformer] update bloom/llama/vit/chatglm tests
    flybird11111 authored Aug 14, 2023
    Configuration menu
    Copy the full SHA
    82ea190 View commit details
    Browse the repository at this point in the history
  4. Merge pull request hpcaitech#4424 from ver217/sync/pipeline

    [sync] update pipeline branch with main
    FrankLeeeee authored Aug 14, 2023
    Configuration menu
    Copy the full SHA
    60db2cc View commit details
    Browse the repository at this point in the history
  5. Configuration menu
    Copy the full SHA
    9d1a6d2 View commit details
    Browse the repository at this point in the history
  6. Configuration menu
    Copy the full SHA
    4f095e6 View commit details
    Browse the repository at this point in the history