Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug] 仅支持了GShard模式的MoE模型转huggingface #279

Open
Cerberous opened this issue Jul 16, 2024 · 0 comments
Open

[Bug] 仅支持了GShard模式的MoE模型转huggingface #279

Cerberous opened this issue Jul 16, 2024 · 0 comments
Assignees
Labels
bug Something isn't working

Comments

@Cerberous
Copy link

Describe the bug

  1. 之前给出的脚本仅支持了GShard MoE训练的方式转化hf的脚本,但是如果用MegaBlock进行训练的话权重转换脚本就不适用了。
  2. 仍然未提供已经训练好的Internevo的权重转换成internevo MoE权重的脚本。

Environment

官方镜像

Other information

No response

@Cerberous Cerberous added the bug Something isn't working label Jul 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants