RuntimeError: Error(s) in loading state_dict for ActionModel:
Missing key(s) in state_dict: "net.blocks.12.attn.qkv.weight", "net.blocks.12.attn.qkv.bias", "net.blocks.12.attn.proj.weight", "net.blocks.12.attn.proj.
bias", "net.blocks.12.mlp.fc1.weight", "net.blocks.12.mlp.fc1.bias", "net.blocks.12.mlp.fc2.weight", "net.blocks.12.mlp.fc2.bias", "net.blocks.13.attn.qkv.weight
", "net.blocks.13.attn.qkv.bias", "net.blocks.13.attn.proj.weight", "net.blocks.13.attn.proj.bias", "net.blocks.13.mlp.fc1.weight", "net.blocks.13.mlp.fc1.bias",
"net.blocks.13.mlp.fc2.weight", "net.blocks.13.mlp.fc2.bias", "net.blocks.14.attn.qkv.weight", "net.blocks.14.attn.qkv.bias", "net.blocks.14.attn.proj.weight",
"net.blocks.14.attn.proj.bias", "net.blocks.14.mlp.fc1.weight", "net.blocks.14.mlp.fc1.bias", "net.blocks.14.mlp.fc2.weight", "net.blocks.14.mlp.fc2.bias", "net.
blocks.15.attn.qkv.weight", "net.blocks.15.attn.qkv.bias", "net.blocks.15.attn.proj.weight", "net.blocks.15.attn.proj.bias", "net.blocks.15.mlp.fc1.weight", "net
.blocks.15.mlp.fc1.bias", "net.blocks.15.mlp.fc2.weight", "net.blocks.15.mlp.fc2.bias", "net.blocks.16.attn.qkv.weight", "net.blocks.16.attn.qkv.bias", "net.bloc
ks.16.attn.proj.weight", "net.blocks.16.attn.proj.bias", "net.blocks.16.mlp.fc1.weight", "net.blocks.16.mlp.fc1.bias", "net.blocks.16.mlp.fc2.weight", "net.block
s.16.mlp.fc2.bias", "net.blocks.17.attn.qkv.weight", "net.blocks.17.attn.qkv.bias", "net.blocks.17.attn.proj.weight", "net.blocks.17.attn.proj.bias", "net.blocks
.17.mlp.fc1.weight", "net.blocks.17.mlp.fc1.bias", "net.blocks.17.mlp.fc2.weight", "net.blocks.17.mlp.fc2.bias", "net.blocks.18.attn.qkv.weight", "net.blocks.18.
attn.qkv.bias", "net.blocks.18.attn.proj.weight", "net.blocks.18.attn.proj.bias", "net.blocks.18.mlp.fc1.weight", "net.blocks.18.mlp.fc1.bias", "net.blocks.18.ml
p.fc2.weight", "net.blocks.18.mlp.fc2.bias", "net.blocks.19.attn.qkv.weight", "net.blocks.19.attn.qkv.bias", "net.blocks.19.attn.proj.weight", "net.blocks.19.att
n.proj.bias", "net.blocks.19.mlp.fc1.weight", "net.blocks.19.mlp.fc1.bias", "net.blocks.19.mlp.fc2.weight", "net.blocks.19.mlp.fc2.bias", "net.blocks.20.attn.qkv
.weight", "net.blocks.20.attn.qkv.bias", "net.blocks.20.attn.proj.weight", "net.blocks.20.attn.proj.bias", "net.blocks.20.mlp.fc1.weight", "net.blocks.20.mlp.fc1
.bias", "net.blocks.20.mlp.fc2.weight", "net.blocks.20.mlp.fc2.bias", "net.blocks.21.attn.qkv.weight", "net.blocks.21.attn.qkv.bias", "net.blocks.21.attn.proj.we
ight", "net.blocks.21.attn.proj.bias", "net.blocks.21.mlp.fc1.weight", "net.blocks.21.mlp.fc1.bias", "net.blocks.21.mlp.fc2.weight", "net.blocks.21.mlp.fc2.bias"
, "net.blocks.22.attn.qkv.weight", "net.blocks.22.attn.qkv.bias", "net.blocks.22.attn.proj.weight", "net.blocks.22.attn.proj.bias", "net.blocks.22.mlp.fc1.weight
", "net.blocks.22.mlp.fc1.bias", "net.blocks.22.mlp.fc2.weight", "net.blocks.22.mlp.fc2.bias", "net.blocks.23.attn.qkv.weight", "net.blocks.23.attn.qkv.bias", "n
et.blocks.23.attn.proj.weight", "net.blocks.23.attn.proj.bias", "net.blocks.23.mlp.fc1.weight", "net.blocks.23.mlp.fc1.bias", "net.blocks.23.mlp.fc2.weight", "ne
t.blocks.23.mlp.fc2.bias".
size mismatch for net.positional_embedding: copying a param with shape torch.Size([17, 768]) from checkpoint, the shape in current model is torch.Size([1
7, 1024]).
size mismatch for net.history_embedder.linear.weight: copying a param with shape torch.Size([768, 7]) from checkpoint, the shape in current model is torc
h.Size([1024, 7]).
size mismatch for net.history_embedder.linear.bias: copying a param with shape torch.Size([768]) from checkpoint, the shape in current model is torch.Siz
e([1024]).
size mismatch for net.x_embedder.linear.weight: copying a param with shape torch.Size([768, 7]) from checkpoint, the shape in current model is torch.Size
([1024, 7]).
size mismatch for net.x_embedder.linear.bias: copying a param with shape torch.Size([768]) from checkpoint, the shape in current model is torch.Size([102
4]).
size mismatch for net.t_embedder.mlp.0.weight: copying a param with shape torch.Size([768, 256]) from checkpoint, the shape in current model is torch.Siz
e([1024, 256]).怎么解决
最新发布