Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Philip's blog #31

Open
p208p2002 opened this issue Dec 15, 2023 · 0 comments
Open

Philip's blog #31

p208p2002 opened this issue Dec 15, 2023 · 0 comments

Comments

@p208p2002
Copy link
Owner

https://blog.philip-huang.tech/?page=peft-overview

- tags: peft overview LLM fine-tune LoRA Adapter - date: 2023/12/15

語言模型(LM)技術已經實現一些重大突破,使得模型的規模更加龐大。然而,對大部份的人說,要微調如此巨大的模型所需的門檻太高。Parameter-efficient fine-tuning(PEFT)提供了一種新的訓練方法,即通過訓練一小組參數,使微調門檻降低,並且讓模型能夠適應和執行新的任務。

LM fine-tuning 演進

  1. Full fine-tuning
    Transformer 架構模型剛推出時(BERT,GPT, etc.),普遍模型大小落在500M~700M左右,這時候高端的消費級顯卡可以負擔微調所需的硬體門檻。

  2. In-Context learni

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant