Kohya training doesn't work #1458
Unanswered
loliana232
asked this question in
Q&A
Replies: 1 comment
-
I have the same issue and not sure how to fix it |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello, I tried to run Kohya GUI through Google colab, I used dreambooth training but when I tried to resume training, I give it the folder where the model exists but it gave me this error many times :
RuntimeError: Error(s) in loading state_dict for UNet2DConditionModel:
Missing key(s) in state_dict: "conv_in.weight", "conv_in.bias", "time_embedding.linear_1.weight", "time_embedding.linear_1.bias", "time_embedding.linear_2.weight", "time_embedding.linear_2.bias", "down_blocks.0.attentions.0.norm.weight", "down_blocks.0.attentions.0.norm.bias", "down_blocks.0.attentions.0.proj_in.weight", "down_blocks.0.attentions.0.proj_in.bias", "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.weight", "down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.bias", "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.0.attentions.0.transformer_blocks.0.norm1.weight", "down_blocks.0.attentions.0.transformer_blocks.0.norm1.bias", "down_blocks.0.attentions.0.transformer_blocks.0.norm2.weight", "down_blocks.0.attentions.0.transformer_blocks.0.norm2.bias", "down_blocks.0.attentions.0.transformer_blocks.0.norm3.weight", "down_blocks.0.attentions.0.transformer_blocks.0.norm3.bias", "down_blocks.0.attentions.0.proj_out.weight", "down_blocks.0.attentions.0.proj_out.bias", "down_blocks.0.attentions.1.norm.weight", "down_blocks.0.attentions.1.norm.bias", "down_blocks.0.attentions.1.proj_in.weight", "down_blocks.0.attentions.1.proj_in.bias", "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.weight", "down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.bias", "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.0.attentions.1.transformer_blocks.0.norm1.weight", "down_blocks.0.attentions.1.transformer_blocks.0.norm1.bias", "down_blocks.0.attentions.1.transformer_blocks.0.norm2.weight", "down_blocks.0.attentions.1.transformer_blocks.0.norm2.bias", "down_blocks.0.attentions.1.transformer_blocks.0.norm3.weight", "down_blocks.0.attentions.1.transformer_blocks.0.norm3.bias", "down_blocks.0.attentions.1.proj_out.weight", "down_blocks.0.attentions.1.proj_out.bias", "down_blocks.0.resnets.0.norm1.weight", "down_blocks.0.resnets.0.norm1.bias", "down_blocks.0.resnets.0.conv1.weight", "down_blocks.0.resnets.0.conv1.bias", "down_blocks.0.resnets.0.time_emb_proj.weight", "down_blocks.0.resnets.0.time_emb_proj.bias", "down_blocks.0.resnets.0.norm2.weight", "down_blocks.0.resnets.0.norm2.bias", "down_blocks.0.resnets.0.conv2.weight", "down_blocks.0.resnets.0.conv2.bias", "down_blocks.0.resnets.1.norm1.weight", "down_blocks.0.resnets.1.norm1.bias", "down_blocks.0.resnets.1.conv1.weight", "down_blocks.0.resnets.1.conv1.bias", "down_blocks.0.resnets.1.time_emb_proj.weight", "down_blocks.0.resnets.1.time_emb_proj.bias", "down_blocks.0.resnets.1.norm2.weight", "down_blocks.0.resnets.1.norm2.bias", "down_blocks.0.resnets.1.conv2.weight", "down_blocks.0.resnets.1.conv2.bias", "down_blocks.0.downsamplers.0.conv.weight", "down_blocks.0.downsamplers.0.conv.bias", "down_blocks.1.attentions.0.norm.weight", "down_blocks.1.attentions.0.norm.bias", "down_blocks.1.attentions.0.proj_in.weight", "down_blocks.1.attentions.0.proj_in.bias", "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.weight", "down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.bias", "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.1.attentions.0.transformer_blocks.0.norm1.weight", "down_blocks.1.attentions.0.transformer_blocks.0.norm1.bias", "down_blocks.1.attentions.0.transformer_blocks.0.norm2.weight", "down_blocks.1.attentions.0.transformer_blocks.0.norm2.bias", "down_blocks.1.attentions.0.transformer_blocks.0.norm3.weight", "down_blocks.1.attentions.0.transformer_blocks.0.norm3.bias", "down_blocks.1.attentions.0.proj_out.weight", "down_blocks.1.attentions.0.proj_out.bias", "down_blocks.1.attentions.1.norm.weight", "down_blocks.1.attentions.1.norm.bias", "down_blocks.1.attentions.1.proj_in.weight", "down_blocks.1.attentions.1.proj_in.bias", "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.weight", "down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.bias", "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.1.attentions.1.transformer_blocks.0.norm1.weight", "down_blocks.1.attentions.1.transformer_blocks.0.norm1.bias", "down_blocks.1.attentions.1.transformer_blocks.0.norm2.weight", "down_blocks.1.attentions.1.transformer_blocks.0.norm2.bias", "down_blocks.1.attentions.1.transformer_blocks.0.norm3.weight", "down_blocks.1.attentions.1.transformer_blocks.0.norm3.bias", "down_blocks.1.attentions.1.proj_out.weight", "down_blocks.1.attentions.1.proj_out.bias", "down_blocks.1.resnets.0.norm1.weight", "down_blocks.1.resnets.0.norm1.bias", "down_blocks.1.resnets.0.conv1.weight", "down_blocks.1.resnets.0.conv1.bias", "down_blocks.1.resnets.0.time_emb_proj.weight", "down_blocks.1.resnets.0.time_emb_proj.bias", "down_blocks.1.resnets.0.norm2.weight", "down_blocks.1.resnets.0.norm2.bias", "down_blocks.1.resnets.0.conv2.weight", "down_blocks.1.resnets.0.conv2.bias", "down_blocks.1.resnets.0.conv_shortcut.weight", "down_blocks.1.resnets.0.conv_shortcut.bias", "down_blocks.1.resnets.1.norm1.weight", "down_blocks.1.resnets.1.norm1.bias", "down_blocks.1.resnets.1.conv1.weight", "down_blocks.1.resnets.1.conv1.bias", "down_blocks.1.resnets.1.time_emb_proj.weight", "down_blocks.1.resnets.1.time_emb_proj.bias", "down_blocks.1.resnets.1.norm2.weight", "down_blocks.1.resnets.1.norm2.bias", "down_blocks.1.resnets.1.conv2.weight", "down_blocks.1.resnets.1.conv2.bias", "down_blocks.1.downsamplers.0.conv.weight", "down_blocks.1.downsamplers.0.conv.bias", "down_blocks.2.attentions.0.norm.weight", "down_blocks.2.attentions.0.norm.bias", "down_blocks.2.attentions.0.proj_in.weight", "down_blocks.2.attentions.0.proj_in.bias", "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.weight", "down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.bias", "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.2.attentions.0.transformer_blocks.0.norm1.weight", "down_blocks.2.attentions.0.transformer_blocks.0.norm1.bias", "down_blocks.2.attentions.0.transformer_blocks.0.norm2.weight", "down_blocks.2.attentions.0.transformer_blocks.0.norm2.bias", "down_blocks.2.attentions.0.transformer_blocks.0.norm3.weight", "down_blocks.2.attentions.0.transformer_blocks.0.norm3.bias", "down_blocks.2.attentions.0.proj_out.weight", "down_blocks.2.attentions.0.proj_out.bias", "down_blocks.2.attentions.1.norm.weight", "down_blocks.2.attentions.1.norm.bias", "down_blocks.2.attentions.1.proj_in.weight", "down_blocks.2.attentions.1.proj_in.bias", "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight", "down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias", "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "down_blocks.2.attentions.1.transformer_blocks.0.norm1.weight", "down_blocks.2.attentions.1.transformer_blocks.0.norm1.bias", "down_blocks.2.attentions.1.transformer_blocks.0.norm2.weight", "down_blocks.2.attentions.1.transformer_blocks.0.norm2.bias", "down_blocks.2.attentions.1.transformer_blocks.0.norm3.weight", "down_blocks.2.attentions.1.transformer_blocks.0.norm3.bias", "down_blocks.2.attentions.1.proj_out.weight", "down_blocks.2.attentions.1.proj_out.bias", "down_blocks.2.resnets.0.norm1.weight", "down_blocks.2.resnets.0.norm1.bias", "down_blocks.2.resnets.0.conv1.weight", "down_blocks.2.resnets.0.conv1.bias", "down_blocks.2.resnets.0.time_emb_proj.weight", "down_blocks.2.resnets.0.time_emb_proj.bias", "down_blocks.2.resnets.0.norm2.weight", "down_blocks.2.resnets.0.norm2.bias", "down_blocks.2.resnets.0.conv2.weight", "down_blocks.2.resnets.0.conv2.bias", "down_blocks.2.resnets.0.conv_shortcut.weight", "down_blocks.2.resnets.0.conv_shortcut.bias", "down_blocks.2.resnets.1.norm1.weight", "down_blocks.2.resnets.1.norm1.bias", "down_blocks.2.resnets.1.conv1.weight", "down_blocks.2.resnets.1.conv1.bias", "down_blocks.2.resnets.1.time_emb_proj.weight", "down_blocks.2.resnets.1.time_emb_proj.bias", "down_blocks.2.resnets.1.norm2.weight", "down_blocks.2.resnets.1.norm2.bias", "down_blocks.2.resnets.1.conv2.weight", "down_blocks.2.resnets.1.conv2.bias", "down_blocks.2.downsamplers.0.conv.weight", "down_blocks.2.downsamplers.0.conv.bias", "down_blocks.3.resnets.0.norm1.weight", "down_blocks.3.resnets.0.norm1.bias", "down_blocks.3.resnets.0.conv1.weight", "down_blocks.3.resnets.0.conv1.bias", "down_blocks.3.resnets.0.time_emb_proj.weight", "down_blocks.3.resnets.0.time_emb_proj.bias", "down_blocks.3.resnets.0.norm2.weight", "down_blocks.3.resnets.0.norm2.bias", "down_blocks.3.resnets.0.conv2.weight", "down_blocks.3.resnets.0.conv2.bias", "down_blocks.3.resnets.1.norm1.weight", "down_blocks.3.resnets.1.norm1.bias", "down_blocks.3.resnets.1.conv1.weight", "down_blocks.3.resnets.1.conv1.bias", "down_blocks.3.resnets.1.time_emb_proj.weight", "down_blocks.3.resnets.1.time_emb_proj.bias", "down_blocks.3.resnets.1.norm2.weight", "down_blocks.3.resnets.1.norm2.bias", "down_blocks.3.resnets.1.conv2.weight", "down_blocks.3.resnets.1.conv2.bias", "up_blocks.0.resnets.0.norm1.weight", "up_blocks.0.resnets.0.norm1.bias", "up_blocks.0.resnets.0.conv1.weight", "up_blocks.0.resnets.0.conv1.bias", "up_blocks.0.resnets.0.time_emb_proj.weight", "up_blocks.0.resnets.0.time_emb_proj.bias", "up_blocks.0.resnets.0.norm2.weight", "up_blocks.0.resnets.0.norm2.bias", "up_blocks.0.resnets.0.conv2.weight", "up_blocks.0.resnets.0.conv2.bias", "up_blocks.0.resnets.0.conv_shortcut.weight", "up_blocks.0.resnets.0.conv_shortcut.bias", "up_blocks.0.resnets.1.norm1.weight", "up_blocks.0.resnets.1.norm1.bias", "up_blocks.0.resnets.1.conv1.weight", "up_blocks.0.resnets.1.conv1.bias", "up_blocks.0.resnets.1.time_emb_proj.weight", "up_blocks.0.resnets.1.time_emb_proj.bias", "up_blocks.0.resnets.1.norm2.weight", "up_blocks.0.resnets.1.norm2.bias", "up_blocks.0.resnets.1.conv2.weight", "up_blocks.0.resnets.1.conv2.bias", "up_blocks.0.resnets.1.conv_shortcut.weight", "up_blocks.0.resnets.1.conv_shortcut.bias", "up_blocks.0.resnets.2.norm1.weight", "up_blocks.0.resnets.2.norm1.bias", "up_blocks.0.resnets.2.conv1.weight", "up_blocks.0.resnets.2.conv1.bias", "up_blocks.0.resnets.2.time_emb_proj.weight", "up_blocks.0.resnets.2.time_emb_proj.bias", "up_blocks.0.resnets.2.norm2.weight", "up_blocks.0.resnets.2.norm2.bias", "up_blocks.0.resnets.2.conv2.weight", "up_blocks.0.resnets.2.conv2.bias", "up_blocks.0.resnets.2.conv_shortcut.weight", "up_blocks.0.resnets.2.conv_shortcut.bias", "up_blocks.0.upsamplers.0.conv.weight", "up_blocks.0.upsamplers.0.conv.bias", "up_blocks.1.attentions.0.norm.weight", "up_blocks.1.attentions.0.norm.bias", "up_blocks.1.attentions.0.proj_in.weight", "up_blocks.1.attentions.0.proj_in.bias", "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.weight", "up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.bias", "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.1.attentions.0.transformer_blocks.0.norm1.weight", "up_blocks.1.attentions.0.transformer_blocks.0.norm1.bias", "up_blocks.1.attentions.0.transformer_blocks.0.norm2.weight", "up_blocks.1.attentions.0.transformer_blocks.0.norm2.bias", "up_blocks.1.attentions.0.transformer_blocks.0.norm3.weight", "up_blocks.1.attentions.0.transformer_blocks.0.norm3.bias", "up_blocks.1.attentions.0.proj_out.weight", "up_blocks.1.attentions.0.proj_out.bias", "up_blocks.1.attentions.1.norm.weight", "up_blocks.1.attentions.1.norm.bias", "up_blocks.1.attentions.1.proj_in.weight", "up_blocks.1.attentions.1.proj_in.bias", "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.weight", "up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.bias", "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.1.attentions.1.transformer_blocks.0.norm1.weight", "up_blocks.1.attentions.1.transformer_blocks.0.norm1.bias", "up_blocks.1.attentions.1.transformer_blocks.0.norm2.weight", "up_blocks.1.attentions.1.transformer_blocks.0.norm2.bias", "up_blocks.1.attentions.1.transformer_blocks.0.norm3.weight", "up_blocks.1.attentions.1.transformer_blocks.0.norm3.bias", "up_blocks.1.attentions.1.proj_out.weight", "up_blocks.1.attentions.1.proj_out.bias", "up_blocks.1.attentions.2.norm.weight", "up_blocks.1.attentions.2.norm.bias", "up_blocks.1.attentions.2.proj_in.weight", "up_blocks.1.attentions.2.proj_in.bias", "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.weight", "up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.bias", "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.1.attentions.2.transformer_blocks.0.norm1.weight", "up_blocks.1.attentions.2.transformer_blocks.0.norm1.bias", "up_blocks.1.attentions.2.transformer_blocks.0.norm2.weight", "up_blocks.1.attentions.2.transformer_blocks.0.norm2.bias", "up_blocks.1.attentions.2.transformer_blocks.0.norm3.weight", "up_blocks.1.attentions.2.transformer_blocks.0.norm3.bias", "up_blocks.1.attentions.2.proj_out.weight", "up_blocks.1.attentions.2.proj_out.bias", "up_blocks.1.resnets.0.norm1.weight", "up_blocks.1.resnets.0.norm1.bias", "up_blocks.1.resnets.0.conv1.weight", "up_blocks.1.resnets.0.conv1.bias", "up_blocks.1.resnets.0.time_emb_proj.weight", "up_blocks.1.resnets.0.time_emb_proj.bias", "up_blocks.1.resnets.0.norm2.weight", "up_blocks.1.resnets.0.norm2.bias", "up_blocks.1.resnets.0.conv2.weight", "up_blocks.1.resnets.0.conv2.bias", "up_blocks.1.resnets.0.conv_shortcut.weight", "up_blocks.1.resnets.0.conv_shortcut.bias", "up_blocks.1.resnets.1.norm1.weight", "up_blocks.1.resnets.1.norm1.bias", "up_blocks.1.resnets.1.conv1.weight", "up_blocks.1.resnets.1.conv1.bias", "up_blocks.1.resnets.1.time_emb_proj.weight", "up_blocks.1.resnets.1.time_emb_proj.bias", "up_blocks.1.resnets.1.norm2.weight", "up_blocks.1.resnets.1.norm2.bias", "up_blocks.1.resnets.1.conv2.weight", "up_blocks.1.resnets.1.conv2.bias", "up_blocks.1.resnets.1.conv_shortcut.weight", "up_blocks.1.resnets.1.conv_shortcut.bias", "up_blocks.1.resnets.2.norm1.weight", "up_blocks.1.resnets.2.norm1.bias", "up_blocks.1.resnets.2.conv1.weight", "up_blocks.1.resnets.2.conv1.bias", "up_blocks.1.resnets.2.time_emb_proj.weight", "up_blocks.1.resnets.2.time_emb_proj.bias", "up_blocks.1.resnets.2.norm2.weight", "up_blocks.1.resnets.2.norm2.bias", "up_blocks.1.resnets.2.conv2.weight", "up_blocks.1.resnets.2.conv2.bias", "up_blocks.1.resnets.2.conv_shortcut.weight", "up_blocks.1.resnets.2.conv_shortcut.bias", "up_blocks.1.upsamplers.0.conv.weight", "up_blocks.1.upsamplers.0.conv.bias", "up_blocks.2.attentions.0.norm.weight", "up_blocks.2.attentions.0.norm.bias", "up_blocks.2.attentions.0.proj_in.weight", "up_blocks.2.attentions.0.proj_in.bias", "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.weight", "up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.bias", "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.2.attentions.0.transformer_blocks.0.norm1.weight", "up_blocks.2.attentions.0.transformer_blocks.0.norm1.bias", "up_blocks.2.attentions.0.transformer_blocks.0.norm2.weight", "up_blocks.2.attentions.0.transformer_blocks.0.norm2.bias", "up_blocks.2.attentions.0.transformer_blocks.0.norm3.weight", "up_blocks.2.attentions.0.transformer_blocks.0.norm3.bias", "up_blocks.2.attentions.0.proj_out.weight", "up_blocks.2.attentions.0.proj_out.bias", "up_blocks.2.attentions.1.norm.weight", "up_blocks.2.attentions.1.norm.bias", "up_blocks.2.attentions.1.proj_in.weight", "up_blocks.2.attentions.1.proj_in.bias", "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight", "up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias", "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.2.attentions.1.transformer_blocks.0.norm1.weight", "up_blocks.2.attentions.1.transformer_blocks.0.norm1.bias", "up_blocks.2.attentions.1.transformer_blocks.0.norm2.weight", "up_blocks.2.attentions.1.transformer_blocks.0.norm2.bias", "up_blocks.2.attentions.1.transformer_blocks.0.norm3.weight", "up_blocks.2.attentions.1.transformer_blocks.0.norm3.bias", "up_blocks.2.attentions.1.proj_out.weight", "up_blocks.2.attentions.1.proj_out.bias", "up_blocks.2.attentions.2.norm.weight", "up_blocks.2.attentions.2.norm.bias", "up_blocks.2.attentions.2.proj_in.weight", "up_blocks.2.attentions.2.proj_in.bias", "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2.weight", "up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2.bias", "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.2.attentions.2.transformer_blocks.0.norm1.weight", "up_blocks.2.attentions.2.transformer_blocks.0.norm1.bias", "up_blocks.2.attentions.2.transformer_blocks.0.norm2.weight", "up_blocks.2.attentions.2.transformer_blocks.0.norm2.bias", "up_blocks.2.attentions.2.transformer_blocks.0.norm3.weight", "up_blocks.2.attentions.2.transformer_blocks.0.norm3.bias", "up_blocks.2.attentions.2.proj_out.weight", "up_blocks.2.attentions.2.proj_out.bias", "up_blocks.2.resnets.0.norm1.weight", "up_blocks.2.resnets.0.norm1.bias", "up_blocks.2.resnets.0.conv1.weight", "up_blocks.2.resnets.0.conv1.bias", "up_blocks.2.resnets.0.time_emb_proj.weight", "up_blocks.2.resnets.0.time_emb_proj.bias", "up_blocks.2.resnets.0.norm2.weight", "up_blocks.2.resnets.0.norm2.bias", "up_blocks.2.resnets.0.conv2.weight", "up_blocks.2.resnets.0.conv2.bias", "up_blocks.2.resnets.0.conv_shortcut.weight", "up_blocks.2.resnets.0.conv_shortcut.bias", "up_blocks.2.resnets.1.norm1.weight", "up_blocks.2.resnets.1.norm1.bias", "up_blocks.2.resnets.1.conv1.weight", "up_blocks.2.resnets.1.conv1.bias", "up_blocks.2.resnets.1.time_emb_proj.weight", "up_blocks.2.resnets.1.time_emb_proj.bias", "up_blocks.2.resnets.1.norm2.weight", "up_blocks.2.resnets.1.norm2.bias", "up_blocks.2.resnets.1.conv2.weight", "up_blocks.2.resnets.1.conv2.bias", "up_blocks.2.resnets.1.conv_shortcut.weight", "up_blocks.2.resnets.1.conv_shortcut.bias", "up_blocks.2.resnets.2.norm1.weight", "up_blocks.2.resnets.2.norm1.bias", "up_blocks.2.resnets.2.conv1.weight", "up_blocks.2.resnets.2.conv1.bias", "up_blocks.2.resnets.2.time_emb_proj.weight", "up_blocks.2.resnets.2.time_emb_proj.bias", "up_blocks.2.resnets.2.norm2.weight", "up_blocks.2.resnets.2.norm2.bias", "up_blocks.2.resnets.2.conv2.weight", "up_blocks.2.resnets.2.conv2.bias", "up_blocks.2.resnets.2.conv_shortcut.weight", "up_blocks.2.resnets.2.conv_shortcut.bias", "up_blocks.2.upsamplers.0.conv.weight", "up_blocks.2.upsamplers.0.conv.bias", "up_blocks.3.attentions.0.norm.weight", "up_blocks.3.attentions.0.norm.bias", "up_blocks.3.attentions.0.proj_in.weight", "up_blocks.3.attentions.0.proj_in.bias", "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2.weight", "up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2.bias", "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.3.attentions.0.transformer_blocks.0.norm1.weight", "up_blocks.3.attentions.0.transformer_blocks.0.norm1.bias", "up_blocks.3.attentions.0.transformer_blocks.0.norm2.weight", "up_blocks.3.attentions.0.transformer_blocks.0.norm2.bias", "up_blocks.3.attentions.0.transformer_blocks.0.norm3.weight", "up_blocks.3.attentions.0.transformer_blocks.0.norm3.bias", "up_blocks.3.attentions.0.proj_out.weight", "up_blocks.3.attentions.0.proj_out.bias", "up_blocks.3.attentions.1.norm.weight", "up_blocks.3.attentions.1.norm.bias", "up_blocks.3.attentions.1.proj_in.weight", "up_blocks.3.attentions.1.proj_in.bias", "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2.weight", "up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2.bias", "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.3.attentions.1.transformer_blocks.0.norm1.weight", "up_blocks.3.attentions.1.transformer_blocks.0.norm1.bias", "up_blocks.3.attentions.1.transformer_blocks.0.norm2.weight", "up_blocks.3.attentions.1.transformer_blocks.0.norm2.bias", "up_blocks.3.attentions.1.transformer_blocks.0.norm3.weight", "up_blocks.3.attentions.1.transformer_blocks.0.norm3.bias", "up_blocks.3.attentions.1.proj_out.weight", "up_blocks.3.attentions.1.proj_out.bias", "up_blocks.3.attentions.2.norm.weight", "up_blocks.3.attentions.2.norm.bias", "up_blocks.3.attentions.2.proj_in.weight", "up_blocks.3.attentions.2.proj_in.bias", "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.bias", "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj.weight", "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj.bias", "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2.weight", "up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2.bias", "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.weight", "up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.bias", "up_blocks.3.attentions.2.transformer_blocks.0.norm1.weight", "up_blocks.3.attentions.2.transformer_blocks.0.norm1.bias", "up_blocks.3.attentions.2.transformer_blocks.0.norm2.weight", "up_blocks.3.attentions.2.transformer_blocks.0.norm2.bias", "up_blocks.3.attentions.2.transformer_blocks.0.norm3.weight", "up_blocks.3.attentions.2.transformer_blocks.0.norm3.bias", "up_blocks.3.attentions.2.proj_out.weight", "up_blocks.3.attentions.2.proj_out.bias", "up_blocks.3.resnets.0.norm1.weight", "up_blocks.3.resnets.0.norm1.bias", "up_blocks.3.resnets.0.conv1.weight", "up_blocks.3.resnets.0.conv1.bias", "up_blocks.3.resnets.0.time_emb_proj.weight", "up_blocks.3.resnets.0.time_emb_proj.bias", "up_blocks.3.resnets.0.norm2.weight", "up_blocks.3.resnets.0.norm2.bias", "up_blocks.3.resnets.0.conv2.weight", "up_blocks.3.resnets.0.conv2.bias", "up_blocks.3.resnets.0.conv_shortcut.weight", "up_blocks.3.resnets.0.conv_shortcut.bias", "up_blocks.3.resnets.1.norm1.weight", "up_blocks.3.resnets.1.norm1.bias", "up_blocks.3.resnets.1.conv1.weight", "up_blocks.3.resnets.1.conv1.bias", "up_blocks.3.resnets.1.time_emb_proj.weight", "up_blocks.3.resnets.1.time_emb_proj.bias", "up_blocks.3.resnets.1.norm2.weight", "up_blocks.3.resnets.1.norm2.bias", "up_blocks.3.resnets.1.conv2.weight", "up_blocks.3.resnets.1.conv2.bias", "up_blocks.3.resnets.1.conv_shortcut.weight", "up_blocks.3.resnets.1.conv_shortcut.bias", "up_blocks.3.resnets.2.norm1.weight", "up_blocks.3.resnets.2.norm1.bias", "up_blocks.3.resnets.2.conv1.weight", "up_blocks.3.resnets.2.conv1.bias", "up_blocks.3.resnets.2.time_emb_proj.weight", "up_blocks.3.resnets.2.time_emb_proj.bias", "up_blocks.3.resnets.2.norm2.weight", "up_blocks.3.resnets.2.norm2.bias", "up_blocks.3.resnets.2.conv2.weight", "up_blocks.3.resnets.2.conv2.bias", "up_blocks.3.resnets.2.conv_shortcut.weight", "up_blocks.3.resnets.2.conv_shortcut.bias", "mid_block.attentions.0.norm.weight", "mid_block.attentions.0.norm.bias", "mid_block.attentions.0.proj_in.weight", "mid_block.attentions.0.proj_in.bias", "mid_block.attentions.0.transformer_blocks.0.attn1.to_q.weight", "mid_block.attentions.0.transformer_blocks.0.attn1.to_k.weight", "mid_block.attentions.0.transformer_blocks.0.attn1.to_v.weight", "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.weight", "mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.bias", "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.weight", "mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.bias", "mid_block.attentions.0.transformer_blocks.0.ff.net.2.weight", "mid_block.attentions.0.transformer_blocks.0.ff.net.2.bias", "mid_block.attentions.0.transformer_blocks.0.attn2.to_q.weight", "mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight", "mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight", "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.weight", "mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.bias", "mid_block.attentions.0.transformer_blocks.0.norm1.weight", "mid_block.attentions.0.transformer_blocks.0.norm1.bias", "mid_block.attentions.0.transformer_blocks.0.norm2.weight", "mid_block.attentions.0.transformer_blocks.0.norm2.bias", "mid_block.attentions.0.transformer_blocks.0.norm3.weight", "mid_block.attentions.0.transformer_blocks.0.norm3.bias", "mid_block.attentions.0.proj_out.weight", "mid_block.attentions.0.proj_out.bias", "mid_block.resnets.0.norm1.weight", "mid_block.resnets.0.norm1.bias", "mid_block.resnets.0.conv1.weight", "mid_block.resnets.0.conv1.bias", "mid_block.resnets.0.time_emb_proj.weight", "mid_block.resnets.0.time_emb_proj.bias", "mid_block.resnets.0.norm2.weight", "mid_block.resnets.0.norm2.bias", "mid_block.resnets.0.conv2.weight", "mid_block.resnets.0.conv2.bias", "mid_block.resnets.1.norm1.weight", "mid_block.resnets.1.norm1.bias", "mid_block.resnets.1.conv1.weight", "mid_block.resnets.1.conv1.bias", "mid_block.resnets.1.time_emb_proj.weight", "mid_block.resnets.1.time_emb_proj.bias", "mid_block.resnets.1.norm2.weight", "mid_block.resnets.1.norm2.bias", "mid_block.resnets.1.conv2.weight", "mid_block.resnets.1.conv2.bias", "conv_norm_out.weight", "conv_norm_out.bias", "conv_out.weight", "conv_out.bias".
Unexpected key(s) in state_dict: "logit_scale", "text_model.embeddings.position_ids", "text_model.embeddings.token_embedding.weight", "text_model.embeddings.position_embedding.weight", "text_model.encoder.layers.0.self_attn.k_proj.weight", "text_model.encoder.layers.0.self_attn.k_proj.bias", "text_model.encoder.layers.0.self_attn.v_proj.weight", "text_model.encoder.layers.0.self_attn.v_proj.bias", "text_model.encoder.layers.0.self_attn.q_proj.weight", "text_model.encoder.layers.0.self_attn.q_proj.bias", "text_model.encoder.layers.0.self_attn.out_proj.weight", "text_model.encoder.layers.0.self_attn.out_proj.bias", "text_model.encoder.layers.0.layer_norm1.weight", "text_model.encoder.layers.0.layer_norm1.bias", "text_model.encoder.layers.0.mlp.fc1.weight", "text_model.encoder.layers.0.mlp.fc1.bias", "text_model.encoder.layers.0.mlp.fc2.weight", "text_model.encoder.layers.0.mlp.fc2.bias", "text_model.encoder.layers.0.layer_norm2.weight", "text_model.encoder.layers.0.layer_norm2.bias", "text_model.encoder.layers.1.self_attn.k_proj.weight", "text_model.encoder.layers.1.self_attn.k_proj.bias", "text_model.encoder.layers.1.self_attn.v_proj.weight", "text_model.encoder.layers.1.self_attn.v_proj.bias", "text_model.encoder.layers.1.self_attn.q_proj.weight", "text_model.encoder.layers.1.self_attn.q_proj.bias", "text_model.encoder.layers.1.self_attn.out_proj.weight", "text_model.encoder.layers.1.self_attn.out_proj.bias", "text_model.encoder.layers.1.layer_norm1.weight", "text_model.encoder.layers.1.layer_norm1.bias", "text_model.encoder.layers.1.mlp.fc1.weight", "text_model.encoder.layers.1.mlp.fc1.bias", "text_model.encoder.layers.1.mlp.fc2.weight", "text_model.encoder.layers.1.mlp.fc2.bias", "text_model.encoder.layers.1.layer_norm2.weight", "text_model.encoder.layers.1.layer_norm2.bias", "text_model.encoder.layers.2.self_attn.k_proj.weight", "text_model.encoder.layers.2.self_attn.k_proj.bias", "text_model.encoder.layers.2.self_attn.v_proj.weight", "text_model.encoder.layers.2.self_attn.v_proj.bias", "text_model.encoder.layers.2.self_attn.q_proj.weight", "text_model.encoder.layers.2.self_attn.q_proj.bias", "text_model.encoder.layers.2.self_attn.out_proj.weight", "text_model.encoder.layers.2.self_attn.out_proj.bias", "text_model.encoder.layers.2.layer_norm1.weight", "text_model.encoder.layers.2.layer_norm1.bias", "text_model.encoder.layers.2.mlp.fc1.weight", "text_model.encoder.layers.2.mlp.fc1.bias", "text_model.encoder.layers.2.mlp.fc2.weight", "text_model.encoder.layers.2.mlp.fc2.bias", "text_model.encoder.layers.2.layer_norm2.weight", "text_model.encoder.layers.2.layer_norm2.bias", "text_model.encoder.layers.3.self_attn.k_proj.weight", "text_model.encoder.layers.3.self_attn.k_proj.bias", "text_model.encoder.layers.3.self_attn.v_proj.weight", "text_model.encoder.layers.3.self_attn.v_proj.bias", "text_model.encoder.layers.3.self_attn.q_proj.weight", "text_model.encoder.layers.3.self_attn.q_proj.bias", "text_model.encoder.layers.3.self_attn.out_proj.weight", "text_model.encoder.layers.3.self_attn.out_proj.bias", "text_model.encoder.layers.3.layer_norm1.weight", "text_model.encoder.layers.3.layer_norm1.bias", "text_model.encoder.layers.3.mlp.fc1.weight", "text_model.encoder.layers.3.mlp.fc1.bias", "text_model.encoder.layers.3.mlp.fc2.weight", "text_model.encoder.layers.3.mlp.fc2.bias", "text_model.encoder.layers.3.layer_norm2.weight", "text_model.encoder.layers.3.layer_norm2.bias", "text_model.encoder.layers.4.self_attn.k_proj.weight", "text_model.encoder.layers.4.self_attn.k_proj.bias", "text_model.encoder.layers.4.self_attn.v_proj.weight", "text_model.encoder.layers.4.self_attn.v_proj.bias", "text_model.encoder.layers.4.self_attn.q_proj.weight", "text_model.encoder.layers.4.self_attn.q_proj.bias", "text_model.encoder.layers.4.self_attn.out_proj.weight", "text_model.encoder.layers.4.self_attn.out_proj.bias", "text_model.encoder.layers.4.layer_norm1.weight", "text_model.encoder.layers.4.layer_norm1.bias", "text_model.encoder.layers.4.mlp.fc1.weight", "text_model.encoder.layers.4.mlp.fc1.bias", "text_model.encoder.layers.4.mlp.fc2.weight", "text_model.encoder.layers.4.mlp.fc2.bias", "text_model.encoder.layers.4.layer_norm2.weight", "text_model.encoder.layers.4.layer_norm2.bias", "text_model.encoder.layers.5.self_attn.k_proj.weight", "text_model.encoder.layers.5.self_attn.k_proj.bias", "text_model.encoder.layers.5.self_attn.v_proj.weight", "text_model.encoder.layers.5.self_attn.v_proj.bias", "text_model.encoder.layers.5.self_attn.q_proj.weight", "text_model.encoder.layers.5.self_attn.q_proj.bias", "text_model.encoder.layers.5.self_attn.out_proj.weight", "text_model.encoder.layers.5.self_attn.out_proj.bias", "text_model.encoder.layers.5.layer_norm1.weight", "text_model.encoder.layers.5.layer_norm1.bias", "text_model.encoder.layers.5.mlp.fc1.weight", "text_model.encoder.layers.5.mlp.fc1.bias", "text_model.encoder.layers.5.mlp.fc2.weight", "text_model.encoder.layers.5.mlp.fc2.bias", "text_model.encoder.layers.5.layer_norm2.weight", "text_model.encoder.layers.5.layer_norm2.bias", "text_model.encoder.layers.6.self_attn.k_proj.weight", "text_model.encoder.layers.6.self_attn.k_proj.bias", "text_model.encoder.layers.6.self_attn.v_proj.weight", "text_model.encoder.layers.6.self_attn.v_proj.bias", "text_model.encoder.layers.6.self_attn.q_proj.weight", "text_model.encoder.layers.6.self_attn.q_proj.bias", "text_model.encoder.layers.6.self_attn.out_proj.weight", "text_model.encoder.layers.6.self_attn.out_proj.bias", "text_model.encoder.layers.6.layer_norm1.weight", "text_model.encoder.layers.6.layer_norm1.bias", "text_model.encoder.layers.6.mlp.fc1.weight", "text_model.encoder.layers.6.mlp.fc1.bias", "text_model.encoder.layers.6.mlp.fc2.weight", "text_model.encoder.layers.6.mlp.fc2.bias", "text_model.encoder.layers.6.layer_norm2.weight", "text_model.encoder.layers.6.layer_norm2.bias", "text_model.encoder.layers.7.self_attn.k_proj.weight", "text_model.encoder.layers.7.self_attn.k_proj.bias", "text_model.encoder.layers.7.self_attn.v_proj.weight", "text_model.encoder.layers.7.self_attn.v_proj.bias", "text_model.encoder.layers.7.self_attn.q_proj.weight", "text_model.encoder.layers.7.self_attn.q_proj.bias", "text_model.encoder.layers.7.self_attn.out_proj.weight", "text_model.encoder.layers.7.self_attn.out_proj.bias", "text_model.encoder.layers.7.layer_norm1.weight", "text_model.encoder.layers.7.layer_norm1.bias", "text_model.encoder.layers.7.mlp.fc1.weight", "text_model.encoder.layers.7.mlp.fc1.bias", "text_model.encoder.layers.7.mlp.fc2.weight", "text_model.encoder.layers.7.mlp.fc2.bias", "text_model.encoder.layers.7.layer_norm2.weight", "text_model.encoder.layers.7.layer_norm2.bias", "text_model.encoder.layers.8.self_attn.k_proj.weight", "text_model.encoder.layers.8.self_attn.k_proj.bias", "text_model.encoder.layers.8.self_attn.v_proj.weight", "text_model.encoder.layers.8.self_attn.v_proj.bias", "text_model.encoder.layers.8.self_attn.q_proj.weight", "text_model.encoder.layers.8.self_attn.q_proj.bias", "text_model.encoder.layers.8.self_attn.out_proj.weight", "text_model.encoder.layers.8.self_attn.out_proj.bias", "text_model.encoder.layers.8.layer_norm1.weight", "text_model.encoder.layers.8.layer_norm1.bias", "text_model.encoder.layers.8.mlp.fc1.weight", "text_model.encoder.layers.8.mlp.fc1.bias", "text_model.encoder.layers.8.mlp.fc2.weight", "text_model.encoder.layers.8.mlp.fc2.bias", "text_model.encoder.layers.8.layer_norm2.weight", "text_model.encoder.layers.8.layer_norm2.bias", "text_model.encoder.layers.9.self_attn.k_proj.weight", "text_model.encoder.layers.9.self_attn.k_proj.bias", "text_model.encoder.layers.9.self_attn.v_proj.weight", "text_model.encoder.layers.9.self_attn.v_proj.bias", "text_model.encoder.layers.9.self_attn.q_proj.weight", "text_model.encoder.layers.9.self_attn.q_proj.bias", "text_model.encoder.layers.9.self_attn.out_proj.weight", "text_model.encoder.layers.9.self_attn.out_proj.bias", "text_model.encoder.layers.9.layer_norm1.weight", "text_model.encoder.layers.9.layer_norm1.bias", "text_model.encoder.layers.9.mlp.fc1.weight", "text_model.encoder.layers.9.mlp.fc1.bias", "text_model.encoder.layers.9.mlp.fc2.weight", "text_model.encoder.layers.9.mlp.fc2.bias", "text_model.encoder.layers.9.layer_norm2.weight", "text_model.encoder.layers.9.layer_norm2.bias", "text_model.encoder.layers.10.self_attn.k_proj.weight", "text_model.encoder.layers.10.self_attn.k_proj.bias", "text_model.encoder.layers.10.self_attn.v_proj.weight", "text_model.encoder.layers.10.self_attn.v_proj.bias", "text_model.encoder.layers.10.self_attn.q_proj.weight", "text_model.encoder.layers.10.self_attn.q_proj.bias", "text_model.encoder.layers.10.self_attn.out_proj.weight", "text_model.encoder.layers.10.self_attn.out_proj.bias", "text_model.encoder.layers.10.layer_norm1.weight", "text_model.encoder.layers.10.layer_norm1.bias", "text_model.encoder.layers.10.mlp.fc1.weight", "text_model.encoder.layers.10.mlp.fc1.bias", "text_model.encoder.layers.10.mlp.fc2.weight", "text_model.encoder.layers.10.mlp.fc2.bias", "text_model.encoder.layers.10.layer_norm2.weight", "text_model.encoder.layers.10.layer_norm2.bias", "text_model.encoder.layers.11.self_attn.k_proj.weight", "text_model.encoder.layers.11.self_attn.k_proj.bias", "text_model.encoder.layers.11.self_attn.v_proj.weight", "text_model.encoder.layers.11.self_attn.v_proj.bias", "text_model.encoder.layers.11.self_attn.q_proj.weight", "text_model.encoder.layers.11.self_attn.q_proj.bias", "text_model.encoder.layers.11.self_attn.out_proj.weight", "text_model.encoder.layers.11.self_attn.out_proj.bias", "text_model.encoder.layers.11.layer_norm1.weight", "text_model.encoder.layers.11.layer_norm1.bias", "text_model.encoder.layers.11.mlp.fc1.weight", "text_model.encoder.layers.11.mlp.fc1.bias", "text_model.encoder.layers.11.mlp.fc2.weight", "text_model.encoder.layers.11.mlp.fc2.bias", "text_model.encoder.layers.11.layer_norm2.weight", "text_model.encoder.layers.11.layer_norm2.bias", "text_model.final_layer_norm.weight", "text_model.final_layer_norm.bias", "vision_model.embeddings.class_embedding", "vision_model.embeddings.position_ids", "vision_model.embeddings.patch_embedding.weight", "vision_model.embeddings.position_embedding.weight", "vision_model.pre_layrnorm.weight", "vision_model.pre_layrnorm.bias", "vision_model.encoder.layers.0.self_attn.k_proj.weight", "vision_model.encoder.layers.0.self_attn.k_proj.bias", "vision_model.encoder.layers.0.self_attn.v_proj.weight", "vision_model.encoder.layers.0.self_attn.v_proj.bias", "vision_model.encoder.layers.0.self_attn.q_proj.weight", "vision_model.encoder.layers.0.self_attn.q_proj.bias", "vision_model.encoder.layers.0.self_attn.out_proj.weight", "vision_model.encoder.layers.0.self_attn.out_proj.bias", "vision_model.encoder.layers.0.layer_norm1.weight", "vision_model.encoder.layers.0.layer_norm1.bias", "vision_model.encoder.layers.0.mlp.fc1.weight", "vision_model.encoder.layers.0.mlp.fc1.bias", "vision_model.encoder.layers.0.mlp.fc2.weight", "vision_model.encoder.layers.0.mlp.fc2.bias", "vision_model.encoder.layers.0.layer_norm2.weight", "vision_model.encoder.layers.0.layer_norm2.bias", "vision_model.encoder.layers.1.self_attn.k_proj.weight", "vision_model.encoder.layers.1.self_attn.k_proj.bias", "vision_model.encoder.layers.1.self_attn.v_proj.weight", "vision_model.encoder.layers.1.self_attn.v_proj.bias", "vision_model.encoder.layers.1.self_attn.q_proj.weight", "vision_model.encoder.layers.1.self_attn.q_proj.bias", "vision_model.encoder.layers.1.self_attn.out_proj.weight", "vision_model.encoder.layers.1.self_attn.out_proj.bias", "vision_model.encoder.layers.1.layer_norm1.weight", "vision_model.encoder.layers.1.layer_norm1.bias", "vision_model.encoder.layers.1.mlp.fc1.weight", "vision_model.encoder.layers.1.mlp.fc1.bias", "vision_model.encoder.layers.1.mlp.fc2.weight", "vision_model.encoder.layers.1.mlp.fc2.bias", "vision_model.encoder.layers.1.layer_norm2.weight", "vision_model.encoder.layers.1.layer_norm2.bias", "vision_model.encoder.layers.2.self_attn.k_proj.weight", "vision_model.encoder.layers.2.self_attn.k_proj.bias", "vision_model.encoder.layers.2.self_attn.v_proj.weight", "vision_model.encoder.layers.2.self_attn.v_proj.bias", "vision_model.encoder.layers.2.self_attn.q_proj.weight", "vision_model.encoder.layers.2.self_attn.q_proj.bias", "vision_model.encoder.layers.2.self_attn.out_proj.weight", "vision_model.encoder.layers.2.self_attn.out_proj.bias", "vision_model.encoder.layers.2.layer_norm1.weight", "vision_model.encoder.layers.2.layer_norm1.bias", "vision_model.encoder.layers.2.mlp.fc1.weight", "vision_model.encoder.layers.2.mlp.fc1.bias", "vision_model.encoder.layers.2.mlp.fc2.weight", "vision_model.encoder.layers.2.mlp.fc2.bias", "vision_model.encoder.layers.2.layer_norm2.weight", "vision_model.encoder.layers.2.layer_norm2.bias", "vision_model.encoder.layers.3.self_attn.k_proj.weight", "vision_model.encoder.layers.3.self_attn.k_proj.bias", "vision_model.encoder.layers.3.self_attn.v_proj.weight", "vision_model.encoder.layers.3.self_attn.v_proj.bias", "vision_model.encoder.layers.3.self_attn.q_proj.weight", "vision_model.encoder.layers.3.self_attn.q_proj.bias", "vision_model.encoder.layers.3.self_attn.out_proj.weight", "vision_model.encoder.layers.3.self_attn.out_proj.bias", "vision_model.encoder.layers.3.layer_norm1.weight", "vision_model.encoder.layers.3.layer_norm1.bias", "vision_model.encoder.layers.3.mlp.fc1.weight", "vision_model.encoder.layers.3.mlp.fc1.bias", "vision_model.encoder.layers.3.mlp.fc2.weight", "vision_model.encoder.layers.3.mlp.fc2.bias", "vision_model.encoder.layers.3.layer_norm2.weight", "vision_model.encoder.layers.3.layer_norm2.bias", "vision_model.encoder.layers.4.self_attn.k_proj.weight", "vision_model.encoder.layers.4.self_attn.k_proj.bias", "vision_model.encoder.layers.4.self_attn.v_proj.weight", "vision_model.encoder.layers.4.self_attn.v_proj.bias", "vision_model.encoder.layers.4.self_attn.q_proj.weight", "vision_model.encoder.layers.4.self_attn.q_proj.bias", "vision_model.encoder.layers.4.self_attn.out_proj.weight", "vision_model.encoder.layers.4.self_attn.out_proj.bias", "vision_model.encoder.layers.4.layer_norm1.weight", "vision_model.encoder.layers.4.layer_norm1.bias", "vision_model.encoder.layers.4.mlp.fc1.weight", "vision_model.encoder.layers.4.mlp.fc1.bias", "vision_model.encoder.layers.4.mlp.fc2.weight", "vision_model.encoder.layers.4.mlp.fc2.bias", "vision_model.encoder.layers.4.layer_norm2.weight", "vision_model.encoder.layers.4.layer_norm2.bias", "vision_model.encoder.layers.5.self_attn.k_proj.weight", "vision_model.encoder.layers.5.self_attn.k_proj.bias", "vision_model.encoder.layers.5.self_attn.v_proj.weight", "vision_model.encoder.layers.5.self_attn.v_proj.bias", "vision_model.encoder.layers.5.self_attn.q_proj.weight", "vision_model.encoder.layers.5.self_attn.q_proj.bias", "vision_model.encoder.layers.5.self_attn.out_proj.weight", "vision_model.encoder.layers.5.self_attn.out_proj.bias", "vision_model.encoder.layers.5.layer_norm1.weight", "vision_model.encoder.layers.5.layer_norm1.bias", "vision_model.encoder.layers.5.mlp.fc1.weight", "vision_model.encoder.layers.5.mlp.fc1.bias", "vision_model.encoder.layers.5.mlp.fc2.weight", "vision_model.encoder.layers.5.mlp.fc2.bias", "vision_model.encoder.layers.5.layer_norm2.weight", "vision_model.encoder.layers.5.layer_norm2.bias", "vision_model.encoder.layers.6.self_attn.k_proj.weight", "vision_model.encoder.layers.6.self_attn.k_proj.bias", "vision_model.encoder.layers.6.self_attn.v_proj.weight", "vision_model.encoder.layers.6.self_attn.v_proj.bias", "vision_model.encoder.layers.6.self_attn.q_proj.weight", "vision_model.encoder.layers.6.self_attn.q_proj.bias", "vision_model.encoder.layers.6.self_attn.out_proj.weight", "vision_model.encoder.layers.6.self_attn.out_proj.bias", "vision_model.encoder.layers.6.layer_norm1.weight", "vision_model.encoder.layers.6.layer_norm1.bias", "vision_model.encoder.layers.6.mlp.fc1.weight", "vision_model.encoder.layers.6.mlp.fc1.bias", "vision_model.encoder.layers.6.mlp.fc2.weight", "vision_model.encoder.layers.6.mlp.fc2.bias", "vision_model.encoder.layers.6.layer_norm2.weight", "vision_model.encoder.layers.6.layer_norm2.bias", "vision_model.encoder.layers.7.self_attn.k_proj.weight", "vision_model.encoder.layers.7.self_attn.k_proj.bias", "vision_model.encoder.layers.7.self_attn.v_proj.weight", "vision_model.encoder.layers.7.self_attn.v_proj.bias", "vision_model.encoder.layers.7.self_attn.q_proj.weight", "vision_model.encoder.layers.7.self_attn.q_proj.bias", "vision_model.encoder.layers.7.self_attn.out_proj.weight", "vision_model.encoder.layers.7.self_attn.out_proj.bias", "vision_model.encoder.layers.7.layer_norm1.weight", "vision_model.encoder.layers.7.layer_norm1.bias", "vision_model.encoder.layers.7.mlp.fc1.weight", "vision_model.encoder.layers.7.mlp.fc1.bias", "vision_model.encoder.layers.7.mlp.fc2.weight", "vision_model.encoder.layers.7.mlp.fc2.bias", "vision_model.encoder.layers.7.layer_norm2.weight", "vision_model.encoder.layers.7.layer_norm2.bias", "vision_model.encoder.layers.8.self_attn.k_proj.weight", "vision_model.encoder.layers.8.self_attn.k_proj.bias", "vision_model.encoder.layers.8.self_attn.v_proj.weight", "vision_model.encoder.layers.8.self_attn.v_proj.bias", "vision_model.encoder.layers.8.self_attn.q_proj.weight", "vision_model.encoder.layers.8.self_attn.q_proj.bias", "vision_model.encoder.layers.8.self_attn.out_proj.weight", "vision_model.encoder.layers.8.self_attn.out_proj.bias", "vision_model.encoder.layers.8.layer_norm1.weight", "vision_model.encoder.layers.8.layer_norm1.bias", "vision_model.encoder.layers.8.mlp.fc1.weight", "vision_model.encoder.layers.8.mlp.fc1.bias", "vision_model.encoder.layers.8.mlp.fc2.weight", "vision_model.encoder.layers.8.mlp.fc2.bias", "vision_model.encoder.layers.8.layer_norm2.weight", "vision_model.encoder.layers.8.layer_norm2.bias", "vision_model.encoder.layers.9.self_attn.k_proj.weight", "vision_model.encoder.layers.9.self_attn.k_proj.bias", "vision_model.encoder.layers.9.self_attn.v_proj.weight", "vision_model.encoder.layers.9.self_attn.v_proj.bias", "vision_model.encoder.layers.9.self_attn.q_proj.weight", "vision_model.encoder.layers.9.self_attn.q_proj.bias", "vision_model.encoder.layers.9.self_attn.out_proj.weight", "vision_model.encoder.layers.9.self_attn.out_proj.bias", "vision_model.encoder.layers.9.layer_norm1.weight", "vision_model.encoder.layers.9.layer_norm1.bias", "vision_model.encoder.layers.9.mlp.fc1.weight", "vision_model.encoder.layers.9.mlp.fc1.bias", "vision_model.encoder.layers.9.mlp.fc2.weight", "vision_model.encoder.layers.9.mlp.fc2.bias", "vision_model.encoder.layers.9.layer_norm2.weight", "vision_model.encoder.layers.9.layer_norm2.bias", "vision_model.encoder.layers.10.self_attn.k_proj.weight", "vision_model.encoder.layers.10.self_attn.k_proj.bias", "vision_model.encoder.layers.10.self_attn.v_proj.weight", "vision_model.encoder.layers.10.self_attn.v_proj.bias", "vision_model.encoder.layers.10.self_attn.q_proj.weight", "vision_model.encoder.layers.10.self_attn.q_proj.bias", "vision_model.encoder.layers.10.self_attn.out_proj.weight", "vision_model.encoder.layers.10.self_attn.out_proj.bias", "vision_model.encoder.layers.10.layer_norm1.weight", "vision_model.encoder.layers.10.layer_norm1.bias", "vision_model.encoder.layers.10.mlp.fc1.weight", "vision_model.encoder.layers.10.mlp.fc1.bias", "vision_model.encoder.layers.10.mlp.fc2.weight", "vision_model.encoder.layers.10.mlp.fc2.bias", "vision_model.encoder.layers.10.layer_norm2.weight", "vision_model.encoder.layers.10.layer_norm2.bias", "vision_model.encoder.layers.11.self_attn.k_proj.weight", "vision_model.encoder.layers.11.self_attn.k_proj.bias", "vision_model.encoder.layers.11.self_attn.v_proj.weight", "vision_model.encoder.layers.11.self_attn.v_proj.bias", "vision_model.encoder.layers.11.self_attn.q_proj.weight", "vision_model.encoder.layers.11.self_attn.q_proj.bias", "vision_model.encoder.layers.11.self_attn.out_proj.weight", "vision_model.encoder.layers.11.self_attn.out_proj.bias", "vision_model.encoder.layers.11.layer_norm1.weight", "vision_model.encoder.layers.11.layer_norm1.bias", "vision_model.encoder.layers.11.mlp.fc1.weight", "vision_model.encoder.layers.11.mlp.fc1.bias", "vision_model.encoder.layers.11.mlp.fc2.weight", "vision_model.encoder.layers.11.mlp.fc2.bias", "vision_model.encoder.layers.11.layer_norm2.weight", "vision_model.encoder.layers.11.layer_norm2.bias", "vision_model.encoder.layers.12.self_attn.k_proj.weight", "vision_model.encoder.layers.12.self_attn.k_proj.bias", "vision_model.encoder.layers.12.self_attn.v_proj.weight", "vision_model.encoder.layers.12.self_attn.v_proj.bias", "vision_model.encoder.layers.12.self_attn.q_proj.weight", "vision_model.encoder.layers.12.self_attn.q_proj.bias", "vision_model.encoder.layers.12.self_attn.out_proj.weight", "vision_model.encoder.layers.12.self_attn.out_proj.bias", "vision_model.encoder.layers.12.layer_norm1.weight", "vision_model.encoder.layers.12.layer_norm1.bias", "vision_model.encoder.layers.12.mlp.fc1.weight", "vision_model.encoder.layers.12.mlp.fc1.bias", "vision_model.encoder.layers.12.mlp.fc2.weight", "vision_model.encoder.layers.12.mlp.fc2.bias", "vision_model.encoder.layers.12.layer_norm2.weight", "vision_model.encoder.layers.12.layer_norm2.bias", "vision_model.encoder.layers.13.self_attn.k_proj.weight", "vision_model.encoder.layers.13.self_attn.k_proj.bias", "vision_model.encoder.layers.13.self_attn.v_proj.weight", "vision_model.encoder.layers.13.self_attn.v_proj.bias", "vision_model.encoder.layers.13.self_attn.q_proj.weight", "vision_model.encoder.layers.13.self_attn.q_proj.bias", "vision_model.encoder.layers.13.self_attn.out_proj.weight", "vision_model.encoder.layers.13.self_attn.out_proj.bias", "vision_model.encoder.layers.13.layer_norm1.weight", "vision_model.encoder.layers.13.layer_norm1.bias", "vision_model.encoder.layers.13.mlp.fc1.weight", "vision_model.encoder.layers.13.mlp.fc1.bias", "vision_model.encoder.layers.13.mlp.fc2.weight", "vision_model.encoder.layers.13.mlp.fc2.bias", "vision_model.encoder.layers.13.layer_norm2.weight", "vision_model.encoder.layers.13.layer_norm2.bias", "vision_model.encoder.layers.14.self_attn.k_proj.weight", "vision_model.encoder.layers.14.self_attn.k_proj.bias", "vision_model.encoder.layers.14.self_attn.v_proj.weight", "vision_model.encoder.layers.14.self_attn.v_proj.bias", "vision_model.encoder.layers.14.self_attn.q_proj.weight", "vision_model.encoder.layers.14.self_attn.q_proj.bias", "vision_model.encoder.layers.14.self_attn.out_proj.weight", "vision_model.encoder.layers.14.self_attn.out_proj.bias", "vision_model.encoder.layers.14.layer_norm1.weight", "vision_model.encoder.layers.14.layer_norm1.bias", "vision_model.encoder.layers.14.mlp.fc1.weight", "vision_model.encoder.layers.14.mlp.fc1.bias", "vision_model.encoder.layers.14.mlp.fc2.weight", "vision_model.encoder.layers.14.mlp.fc2.bias", "vision_model.encoder.layers.14.layer_norm2.weight", "vision_model.encoder.layers.14.layer_norm2.bias", "vision_model.encoder.layers.15.self_attn.k_proj.weight", "vision_model.encoder.layers.15.self_attn.k_proj.bias", "vision_model.encoder.layers.15.self_attn.v_proj.weight", "vision_model.encoder.layers.15.self_attn.v_proj.bias", "vision_model.encoder.layers.15.self_attn.q_proj.weight", "vision_model.encoder.layers.15.self_attn.q_proj.bias", "vision_model.encoder.layers.15.self_attn.out_proj.weight", "vision_model.encoder.layers.15.self_attn.out_proj.bias", "vision_model.encoder.layers.15.layer_norm1.weight", "vision_model.encoder.layers.15.layer_norm1.bias", "vision_model.encoder.layers.15.mlp.fc1.weight", "vision_model.encoder.layers.15.mlp.fc1.bias", "vision_model.encoder.layers.15.mlp.fc2.weight", "vision_model.encoder.layers.15.mlp.fc2.bias", "vision_model.encoder.layers.15.layer_norm2.weight", "vision_model.encoder.layers.15.layer_norm2.bias", "vision_model.encoder.layers.16.self_attn.k_proj.weight", "vision_model.encoder.layers.16.self_attn.k_proj.bias", "vision_model.encoder.layers.16.self_attn.v_proj.weight", "vision_model.encoder.layers.16.self_attn.v_proj.bias", "vision_model.encoder.layers.16.self_attn.q_proj.weight", "vision_model.encoder.layers.16.self_attn.q_proj.bias", "vision_model.encoder.layers.16.self_attn.out_proj.weight", "vision_model.encoder.layers.16.self_attn.out_proj.bias", "vision_model.encoder.layers.16.layer_norm1.weight", "vision_model.encoder.layers.16.layer_norm1.bias", "vision_model.encoder.layers.16.mlp.fc1.weight", "vision_model.encoder.layers.16.mlp.fc1.bias", "vision_model.encoder.layers.16.mlp.fc2.weight", "vision_model.encoder.layers.16.mlp.fc2.bias", "vision_model.encoder.layers.16.layer_norm2.weight", "vision_model.encoder.layers.16.layer_norm2.bias", "vision_model.encoder.layers.17.self_attn.k_proj.weight", "vision_model.encoder.layers.17.self_attn.k_proj.bias", "vision_model.encoder.layers.17.self_attn.v_proj.weight", "vision_model.encoder.layers.17.self_attn.v_proj.bias", "vision_model.encoder.layers.17.self_attn.q_proj.weight", "vision_model.encoder.layers.17.self_attn.q_proj.bias", "vision_model.encoder.layers.17.self_attn.out_proj.weight", "vision_model.encoder.layers.17.self_attn.out_proj.bias", "vision_model.encoder.layers.17.layer_norm1.weight", "vision_model.encoder.layers.17.layer_norm1.bias", "vision_model.encoder.layers.17.mlp.fc1.weight", "vision_model.encoder.layers.17.mlp.fc1.bias", "vision_model.encoder.layers.17.mlp.fc2.weight", "vision_model.encoder.layers.17.mlp.fc2.bias", "vision_model.encoder.layers.17.layer_norm2.weight", "vision_model.encoder.layers.17.layer_norm2.bias", "vision_model.encoder.layers.18.self_attn.k_proj.weight", "vision_model.encoder.layers.18.self_attn.k_proj.bias", "vision_model.encoder.layers.18.self_attn.v_proj.weight", "vision_model.encoder.layers.18.self_attn.v_proj.bias", "vision_model.encoder.layers.18.self_attn.q_proj.weight", "vision_model.encoder.layers.18.self_attn.q_proj.bias", "vision_model.encoder.layers.18.self_attn.out_proj.weight", "vision_model.encoder.layers.18.self_attn.out_proj.bias", "vision_model.encoder.layers.18.layer_norm1.weight", "vision_model.encoder.layers.18.layer_norm1.bias", "vision_model.encoder.layers.18.mlp.fc1.weight", "vision_model.encoder.layers.18.mlp.fc1.bias", "vision_model.encoder.layers.18.mlp.fc2.weight", "vision_model.encoder.layers.18.mlp.fc2.bias", "vision_model.encoder.layers.18.layer_norm2.weight", "vision_model.encoder.layers.18.layer_norm2.bias", "vision_model.encoder.layers.19.self_attn.k_proj.weight", "vision_model.encoder.layers.19.self_attn.k_proj.bias", "vision_model.encoder.layers.19.self_attn.v_proj.weight", "vision_model.encoder.layers.19.self_attn.v_proj.bias", "vision_model.encoder.layers.19.self_attn.q_proj.weight", "vision_model.encoder.layers.19.self_attn.q_proj.bias", "vision_model.encoder.layers.19.self_attn.out_proj.weight", "vision_model.encoder.layers.19.self_attn.out_proj.bias", "vision_model.encoder.layers.19.layer_norm1.weight", "vision_model.encoder.layers.19.layer_norm1.bias", "vision_model.encoder.layers.19.mlp.fc1.weight", "vision_model.encoder.layers.19.mlp.fc1.bias", "vision_model.encoder.layers.19.mlp.fc2.weight", "vision_model.encoder.layers.19.mlp.fc2.bias", "vision_model.encoder.layers.19.layer_norm2.weight", "vision_model.encoder.layers.19.layer_norm2.bias", "vision_model.encoder.layers.20.self_attn.k_proj.weight", "vision_model.encoder.layers.20.self_attn.k_proj.bias", "vision_model.encoder.layers.20.self_attn.v_proj.weight", "vision_model.encoder.layers.20.self_attn.v_proj.bias", "vision_model.encoder.layers.20.self_attn.q_proj.weight", "vision_model.encoder.layers.20.self_attn.q_proj.bias", "vision_model.encoder.layers.20.self_attn.out_proj.weight", "vision_model.encoder.layers.20.self_attn.out_proj.bias", "vision_model.encoder.layers.20.layer_norm1.weight", "vision_model.encoder.layers.20.layer_norm1.bias", "vision_model.encoder.layers.20.mlp.fc1.weight", "vision_model.encoder.layers.20.mlp.fc1.bias", "vision_model.encoder.layers.20.mlp.fc2.weight", "vision_model.encoder.layers.20.mlp.fc2.bias", "vision_model.encoder.layers.20.layer_norm2.weight", "vision_model.encoder.layers.20.layer_norm2.bias", "vision_model.encoder.layers.21.self_attn.k_proj.weight", "vision_model.encoder.layers.21.self_attn.k_proj.bias", "vision_model.encoder.layers.21.self_attn.v_proj.weight", "vision_model.encoder.layers.21.self_attn.v_proj.bias", "vision_model.encoder.layers.21.self_attn.q_proj.weight", "vision_model.encoder.layers.21.self_attn.q_proj.bias", "vision_model.encoder.layers.21.self_attn.out_proj.weight", "vision_model.encoder.layers.21.self_attn.out_proj.bias", "vision_model.encoder.layers.21.layer_norm1.weight", "vision_model.encoder.layers.21.layer_norm1.bias", "vision_model.encoder.layers.21.mlp.fc1.weight", "vision_model.encoder.layers.21.mlp.fc1.bias", "vision_model.encoder.layers.21.mlp.fc2.weight", "vision_model.encoder.layers.21.mlp.fc2.bias", "vision_model.encoder.layers.21.layer_norm2.weight", "vision_model.encoder.layers.21.layer_norm2.bias", "vision_model.encoder.layers.22.self_attn.k_proj.weight", "vision_model.encoder.layers.22.self_attn.k_proj.bias", "vision_model.encoder.layers.22.self_attn.v_proj.weight", "vision_model.encoder.layers.22.self_attn.v_proj.bias", "vision_model.encoder.layers.22.self_attn.q_proj.weight", "vision_model.encoder.layers.22.self_attn.q_proj.bias", "vision_model.encoder.layers.22.self_attn.out_proj.weight", "vision_model.encoder.layers.22.self_attn.out_proj.bias", "vision_model.encoder.layers.22.layer_norm1.weight", "vision_model.encoder.layers.22.layer_norm1.bias", "vision_model.encoder.layers.22.mlp.fc1.weight", "vision_model.encoder.layers.22.mlp.fc1.bias", "vision_model.encoder.layers.22.mlp.fc2.weight", "vision_model.encoder.layers.22.mlp.fc2.bias", "vision_model.encoder.layers.22.layer_norm2.weight", "vision_model.encoder.layers.22.layer_norm2.bias", "vision_model.encoder.layers.23.self_attn.k_proj.weight", "vision_model.encoder.layers.23.self_attn.k_proj.bias", "vision_model.encoder.layers.23.self_attn.v_proj.weight", "vision_model.encoder.layers.23.self_attn.v_proj.bias", "vision_model.encoder.layers.23.self_attn.q_proj.weight", "vision_model.encoder.layers.23.self_attn.q_proj.bias", "vision_model.encoder.layers.23.self_attn.out_proj.weight", "vision_model.encoder.layers.23.self_attn.out_proj.bias", "vision_model.encoder.layers.23.layer_norm1.weight", "vision_model.encoder.layers.23.layer_norm1.bias", "vision_model.encoder.layers.23.mlp.fc1.weight", "vision_model.encoder.layers.23.mlp.fc1.bias", "vision_model.encoder.layers.23.mlp.fc2.weight", "vision_model.encoder.layers.23.mlp.fc2.bias", "vision_model.encoder.layers.23.layer_norm2.weight", "vision_model.encoder.layers.23.layer_norm2.bias", "vision_model.post_layernorm.weight", "vision_model.post_layernorm.bias", "visual_projection.weight", "text_projection.weight".
Traceback (most recent call last):
File "/usr/local/bin/accelerate", line 8, in
sys.exit(main())
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/accelerate_cli.py", line 45, in main
args.func(args)
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py", line 986, in launch_command
simple_launcher(args)
File "/usr/local/lib/python3.10/dist-packages/accelerate/commands/launch.py", line 628, in simple_launcher
raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd)
subprocess.CalledProcessError: Command '['/usr/bin/python3', 'train_db.py', '--enable_bucket', '--pretrained_model_name_or_path=runwayml/stable-diffusion-v1-5', '--train_data_dir=/content/drive/MyDrive/10_Style', '--resolution=768,768', '--output_dir=/content/drive/MyDrive/4_Out', '--save_model_as=safetensors', '--output_name=GPINYH', '--max_data_loader_n_workers=0', '--learning_rate=1e-05', '--lr_scheduler=cosine', '--lr_warmup_steps=878', '--train_batch_size=1', '--max_train_steps=8780', '--save_every_n_epochs=10', '--mixed_precision=fp16', '--save_precision=fp16', '--cache_latents', '--optimizer_type=AdamW8bit', '--max_data_loader_n_workers=0', '--resume=/content/drive/MyDrive/3_Out', '--bucket_reso_steps=64', '--save_state', '--xformers', '--noise_offset=0.05']' returned non-zero exit status 1.
Can anyone help???
Beta Was this translation helpful? Give feedback.
All reactions