Movie / check_ckpt.py
Mudrock's picture
Upload 18 files
4c94b0e
raw
history blame
No virus
56.3 kB
import torch
def keys_in_state_dict(ckpt, device='cpu'):
if device=="cpu":
a = torch.load(ckpt, map_location=torch.device('cpu'))["state_dict"]
else:
a = torch.load(ckpt)["state_dict"]
print("keys_in_state_dict", a.keys())
def check_ckpt_diff(ckpt_a, ckpt_b, key_include=None, key_exclude=None, device='cpu', verbose=True):
if device=="cpu":
a = torch.load(ckpt_a, map_location=torch.device('cpu'))["state_dict"]
b = torch.load(ckpt_b, map_location=torch.device('cpu'))["state_dict"]
else:
a = torch.load(ckpt_a)["state_dict"]
b = torch.load(ckpt_b)["state_dict"]
a_sum = 0
b_sum = 0
difference_count = 0
for k in a.keys():
if key_include is not None and key_include not in k:
continue
if key_exclude is not None and key_exclude in k:
continue
if k in b.keys():
a_sum += torch.sum(a[k])
b_sum += torch.sum(b[k])
if verbose:
if torch.sum(a[k]) != torch.sum(b[k]):
print(f"key {k} is different")
difference_count += 1
print("a_sum: ", a_sum)
print("b_sum: ", b_sum)
print("diff: ", a_sum - b_sum)
if verbose:
print("difference_count: ", difference_count)
return bool(a_sum - b_sum)
# Transformer no freeze:
# check_ckpt_diff("/fsx/clap_logs/2022_09_11-19_37_08-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_11-19_37_08-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.resblocks")
check_ckpt_diff("/fsx/clap_logs/2022_09_29-23_42_40-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_1.pt",
"/fsx/clap_logs/2022_09_29-23_42_40-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_2.pt",
"text_branch.resblocks")
# key module.text_branch.resblocks.0.attn.in_proj_weight is different
# key module.text_branch.resblocks.0.attn.in_proj_bias is different
# key module.text_branch.resblocks.0.attn.out_proj.weight is different
# key module.text_branch.resblocks.0.attn.out_proj.bias is different
# key module.text_branch.resblocks.0.ln_1.weight is different
# key module.text_branch.resblocks.0.ln_1.bias is different
# key module.text_branch.resblocks.0.mlp.c_fc.weight is different
# key module.text_branch.resblocks.0.mlp.c_fc.bias is different
# key module.text_branch.resblocks.0.mlp.c_proj.weight is different
# key module.text_branch.resblocks.0.mlp.c_proj.bias is different
# key module.text_branch.resblocks.0.ln_2.weight is different
# key module.text_branch.resblocks.0.ln_2.bias is different
# key module.text_branch.resblocks.1.attn.in_proj_weight is different
# key module.text_branch.resblocks.1.attn.in_proj_bias is different
# key module.text_branch.resblocks.1.attn.out_proj.weight is different
# key module.text_branch.resblocks.1.attn.out_proj.bias is different
# key module.text_branch.resblocks.1.ln_1.weight is different
# key module.text_branch.resblocks.1.ln_1.bias is different
# key module.text_branch.resblocks.1.mlp.c_fc.weight is different
# key module.text_branch.resblocks.1.mlp.c_fc.bias is different
# key module.text_branch.resblocks.1.mlp.c_proj.weight is different
# key module.text_branch.resblocks.1.mlp.c_proj.bias is different
# key module.text_branch.resblocks.1.ln_2.weight is different
# key module.text_branch.resblocks.1.ln_2.bias is different
# key module.text_branch.resblocks.2.attn.in_proj_weight is different
# key module.text_branch.resblocks.2.attn.in_proj_bias is different
# key module.text_branch.resblocks.2.attn.out_proj.weight is different
# key module.text_branch.resblocks.2.attn.out_proj.bias is different
# key module.text_branch.resblocks.2.ln_1.weight is different
# key module.text_branch.resblocks.2.ln_1.bias is different
# key module.text_branch.resblocks.2.mlp.c_fc.weight is different
# key module.text_branch.resblocks.2.mlp.c_fc.bias is different
# key module.text_branch.resblocks.2.mlp.c_proj.weight is different
# key module.text_branch.resblocks.2.mlp.c_proj.bias is different
# key module.text_branch.resblocks.2.ln_2.weight is different
# key module.text_branch.resblocks.2.ln_2.bias is different
# key module.text_branch.resblocks.3.attn.in_proj_weight is different
# key module.text_branch.resblocks.3.attn.in_proj_bias is different
# key module.text_branch.resblocks.3.attn.out_proj.weight is different
# key module.text_branch.resblocks.3.attn.out_proj.bias is different
# key module.text_branch.resblocks.3.ln_1.weight is different
# key module.text_branch.resblocks.3.ln_1.bias is different
# key module.text_branch.resblocks.3.mlp.c_fc.weight is different
# key module.text_branch.resblocks.3.mlp.c_fc.bias is different
# key module.text_branch.resblocks.3.mlp.c_proj.weight is different
# key module.text_branch.resblocks.3.mlp.c_proj.bias is different
# key module.text_branch.resblocks.3.ln_2.weight is different
# key module.text_branch.resblocks.3.ln_2.bias is different
# key module.text_branch.resblocks.4.attn.in_proj_weight is different
# key module.text_branch.resblocks.4.attn.in_proj_bias is different
# key module.text_branch.resblocks.4.attn.out_proj.weight is different
# key module.text_branch.resblocks.4.attn.out_proj.bias is different
# key module.text_branch.resblocks.4.ln_1.weight is different
# key module.text_branch.resblocks.4.ln_1.bias is different
# key module.text_branch.resblocks.4.mlp.c_fc.weight is different
# key module.text_branch.resblocks.4.mlp.c_fc.bias is different
# key module.text_branch.resblocks.4.mlp.c_proj.weight is different
# key module.text_branch.resblocks.4.mlp.c_proj.bias is different
# key module.text_branch.resblocks.4.ln_2.weight is different
# key module.text_branch.resblocks.4.ln_2.bias is different
# key module.text_branch.resblocks.5.attn.in_proj_weight is different
# key module.text_branch.resblocks.5.attn.in_proj_bias is different
# key module.text_branch.resblocks.5.attn.out_proj.weight is different
# key module.text_branch.resblocks.5.attn.out_proj.bias is different
# key module.text_branch.resblocks.5.ln_1.weight is different
# key module.text_branch.resblocks.5.ln_1.bias is different
# key module.text_branch.resblocks.5.mlp.c_fc.weight is different
# key module.text_branch.resblocks.5.mlp.c_fc.bias is different
# key module.text_branch.resblocks.5.mlp.c_proj.weight is different
# key module.text_branch.resblocks.5.mlp.c_proj.bias is different
# key module.text_branch.resblocks.5.ln_2.weight is different
# key module.text_branch.resblocks.5.ln_2.bias is different
# key module.text_branch.resblocks.6.attn.in_proj_weight is different
# key module.text_branch.resblocks.6.attn.in_proj_bias is different
# key module.text_branch.resblocks.6.attn.out_proj.weight is different
# key module.text_branch.resblocks.6.attn.out_proj.bias is different
# key module.text_branch.resblocks.6.ln_1.weight is different
# key module.text_branch.resblocks.6.ln_1.bias is different
# key module.text_branch.resblocks.6.mlp.c_fc.weight is different
# key module.text_branch.resblocks.6.mlp.c_fc.bias is different
# key module.text_branch.resblocks.6.mlp.c_proj.weight is different
# key module.text_branch.resblocks.6.mlp.c_proj.bias is different
# key module.text_branch.resblocks.6.ln_2.weight is different
# key module.text_branch.resblocks.6.ln_2.bias is different
# key module.text_branch.resblocks.7.attn.in_proj_weight is different
# key module.text_branch.resblocks.7.attn.in_proj_bias is different
# key module.text_branch.resblocks.7.attn.out_proj.weight is different
# key module.text_branch.resblocks.7.attn.out_proj.bias is different
# key module.text_branch.resblocks.7.ln_1.weight is different
# key module.text_branch.resblocks.7.ln_1.bias is different
# key module.text_branch.resblocks.7.mlp.c_fc.weight is different
# key module.text_branch.resblocks.7.mlp.c_fc.bias is different
# key module.text_branch.resblocks.7.mlp.c_proj.weight is different
# key module.text_branch.resblocks.7.mlp.c_proj.bias is different
# key module.text_branch.resblocks.7.ln_2.weight is different
# key module.text_branch.resblocks.7.ln_2.bias is different
# key module.text_branch.resblocks.8.attn.in_proj_weight is different
# key module.text_branch.resblocks.8.attn.in_proj_bias is different
# key module.text_branch.resblocks.8.attn.out_proj.weight is different
# key module.text_branch.resblocks.8.attn.out_proj.bias is different
# key module.text_branch.resblocks.8.ln_1.weight is different
# key module.text_branch.resblocks.8.ln_1.bias is different
# key module.text_branch.resblocks.8.mlp.c_fc.weight is different
# key module.text_branch.resblocks.8.mlp.c_fc.bias is different
# key module.text_branch.resblocks.8.mlp.c_proj.weight is different
# key module.text_branch.resblocks.8.mlp.c_proj.bias is different
# key module.text_branch.resblocks.8.ln_2.weight is different
# key module.text_branch.resblocks.8.ln_2.bias is different
# key module.text_branch.resblocks.9.attn.in_proj_weight is different
# key module.text_branch.resblocks.9.attn.in_proj_bias is different
# key module.text_branch.resblocks.9.attn.out_proj.weight is different
# key module.text_branch.resblocks.9.attn.out_proj.bias is different
# key module.text_branch.resblocks.9.ln_1.weight is different
# key module.text_branch.resblocks.9.ln_1.bias is different
# key module.text_branch.resblocks.9.mlp.c_fc.weight is different
# key module.text_branch.resblocks.9.mlp.c_fc.bias is different
# key module.text_branch.resblocks.9.mlp.c_proj.weight is different
# key module.text_branch.resblocks.9.mlp.c_proj.bias is different
# key module.text_branch.resblocks.9.ln_2.weight is different
# key module.text_branch.resblocks.9.ln_2.bias is different
# key module.text_branch.resblocks.10.attn.in_proj_weight is different
# key module.text_branch.resblocks.10.attn.in_proj_bias is different
# key module.text_branch.resblocks.10.attn.out_proj.weight is different
# key module.text_branch.resblocks.10.attn.out_proj.bias is different
# key module.text_branch.resblocks.10.ln_1.weight is different
# key module.text_branch.resblocks.10.ln_1.bias is different
# key module.text_branch.resblocks.10.mlp.c_fc.weight is different
# key module.text_branch.resblocks.10.mlp.c_fc.bias is different
# key module.text_branch.resblocks.10.mlp.c_proj.weight is different
# key module.text_branch.resblocks.10.mlp.c_proj.bias is different
# key module.text_branch.resblocks.10.ln_2.weight is different
# key module.text_branch.resblocks.10.ln_2.bias is different
# key module.text_branch.resblocks.11.attn.in_proj_weight is different
# key module.text_branch.resblocks.11.attn.in_proj_bias is different
# key module.text_branch.resblocks.11.attn.out_proj.weight is different
# key module.text_branch.resblocks.11.attn.out_proj.bias is different
# key module.text_branch.resblocks.11.ln_1.weight is different
# key module.text_branch.resblocks.11.ln_1.bias is different
# key module.text_branch.resblocks.11.mlp.c_fc.weight is different
# key module.text_branch.resblocks.11.mlp.c_fc.bias is different
# key module.text_branch.resblocks.11.mlp.c_proj.weight is different
# key module.text_branch.resblocks.11.mlp.c_proj.bias is different
# key module.text_branch.resblocks.11.ln_2.weight is different
# key module.text_branch.resblocks.11.ln_2.bias is different
# a_sum: tensor(12113.6445)
# b_sum: tensor(9883.4424)
# diff: tensor(2230.2021)
# True
# Transformer freeze:
# check_ckpt_diff("/fsx/clap_logs/2022_09_16-18_55_10-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_16-18_55_10-model_PANN-14-lr_0.001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.resblocks")
# key module.text_branch.resblocks.0.attn.in_proj_weight is different
# key module.text_branch.resblocks.0.attn.in_proj_bias is different
# key module.text_branch.resblocks.0.attn.out_proj.weight is different
# key module.text_branch.resblocks.0.attn.out_proj.bias is different
# key module.text_branch.resblocks.0.ln_1.weight is different
# key module.text_branch.resblocks.0.ln_1.bias is different
# key module.text_branch.resblocks.0.mlp.c_fc.weight is different
# key module.text_branch.resblocks.0.mlp.c_fc.bias is different
# key module.text_branch.resblocks.0.mlp.c_proj.weight is different
# key module.text_branch.resblocks.0.mlp.c_proj.bias is different
# key module.text_branch.resblocks.0.ln_2.weight is different
# key module.text_branch.resblocks.0.ln_2.bias is different
# key module.text_branch.resblocks.1.attn.in_proj_weight is different
# key module.text_branch.resblocks.1.attn.in_proj_bias is different
# key module.text_branch.resblocks.1.attn.out_proj.weight is different
# key module.text_branch.resblocks.1.attn.out_proj.bias is different
# key module.text_branch.resblocks.1.ln_1.weight is different
# key module.text_branch.resblocks.1.ln_1.bias is different
# key module.text_branch.resblocks.1.mlp.c_fc.weight is different
# key module.text_branch.resblocks.1.mlp.c_fc.bias is different
# key module.text_branch.resblocks.1.mlp.c_proj.weight is different
# key module.text_branch.resblocks.1.mlp.c_proj.bias is different
# key module.text_branch.resblocks.1.ln_2.weight is different
# key module.text_branch.resblocks.1.ln_2.bias is different
# key module.text_branch.resblocks.2.attn.in_proj_weight is different
# key module.text_branch.resblocks.2.attn.in_proj_bias is different
# key module.text_branch.resblocks.2.attn.out_proj.weight is different
# key module.text_branch.resblocks.2.attn.out_proj.bias is different
# key module.text_branch.resblocks.2.ln_1.weight is different
# key module.text_branch.resblocks.2.ln_1.bias is different
# key module.text_branch.resblocks.2.mlp.c_fc.weight is different
# key module.text_branch.resblocks.2.mlp.c_fc.bias is different
# key module.text_branch.resblocks.2.mlp.c_proj.weight is different
# key module.text_branch.resblocks.2.mlp.c_proj.bias is different
# key module.text_branch.resblocks.2.ln_2.weight is different
# key module.text_branch.resblocks.2.ln_2.bias is different
# key module.text_branch.resblocks.3.attn.in_proj_weight is different
# key module.text_branch.resblocks.3.attn.in_proj_bias is different
# key module.text_branch.resblocks.3.attn.out_proj.weight is different
# key module.text_branch.resblocks.3.attn.out_proj.bias is different
# key module.text_branch.resblocks.3.ln_1.weight is different
# key module.text_branch.resblocks.3.ln_1.bias is different
# key module.text_branch.resblocks.3.mlp.c_fc.weight is different
# key module.text_branch.resblocks.3.mlp.c_fc.bias is different
# key module.text_branch.resblocks.3.mlp.c_proj.weight is different
# key module.text_branch.resblocks.3.mlp.c_proj.bias is different
# key module.text_branch.resblocks.3.ln_2.weight is different
# key module.text_branch.resblocks.3.ln_2.bias is different
# key module.text_branch.resblocks.4.attn.in_proj_weight is different
# key module.text_branch.resblocks.4.attn.in_proj_bias is different
# key module.text_branch.resblocks.4.attn.out_proj.weight is different
# key module.text_branch.resblocks.4.attn.out_proj.bias is different
# key module.text_branch.resblocks.4.ln_1.weight is different
# key module.text_branch.resblocks.4.ln_1.bias is different
# key module.text_branch.resblocks.4.mlp.c_fc.weight is different
# key module.text_branch.resblocks.4.mlp.c_fc.bias is different
# key module.text_branch.resblocks.4.mlp.c_proj.weight is different
# key module.text_branch.resblocks.4.mlp.c_proj.bias is different
# key module.text_branch.resblocks.4.ln_2.weight is different
# key module.text_branch.resblocks.4.ln_2.bias is different
# key module.text_branch.resblocks.5.attn.in_proj_weight is different
# key module.text_branch.resblocks.5.attn.in_proj_bias is different
# key module.text_branch.resblocks.5.attn.out_proj.weight is different
# key module.text_branch.resblocks.5.attn.out_proj.bias is different
# key module.text_branch.resblocks.5.ln_1.weight is different
# key module.text_branch.resblocks.5.ln_1.bias is different
# key module.text_branch.resblocks.5.mlp.c_fc.weight is different
# key module.text_branch.resblocks.5.mlp.c_fc.bias is different
# key module.text_branch.resblocks.5.mlp.c_proj.weight is different
# key module.text_branch.resblocks.5.mlp.c_proj.bias is different
# key module.text_branch.resblocks.5.ln_2.weight is different
# key module.text_branch.resblocks.5.ln_2.bias is different
# key module.text_branch.resblocks.6.attn.in_proj_weight is different
# key module.text_branch.resblocks.6.attn.in_proj_bias is different
# key module.text_branch.resblocks.6.attn.out_proj.weight is different
# key module.text_branch.resblocks.6.attn.out_proj.bias is different
# key module.text_branch.resblocks.6.ln_1.weight is different
# key module.text_branch.resblocks.6.ln_1.bias is different
# key module.text_branch.resblocks.6.mlp.c_fc.weight is different
# key module.text_branch.resblocks.6.mlp.c_fc.bias is different
# key module.text_branch.resblocks.6.mlp.c_proj.weight is different
# key module.text_branch.resblocks.6.mlp.c_proj.bias is different
# key module.text_branch.resblocks.6.ln_2.weight is different
# key module.text_branch.resblocks.6.ln_2.bias is different
# key module.text_branch.resblocks.7.attn.in_proj_weight is different
# key module.text_branch.resblocks.7.attn.in_proj_bias is different
# key module.text_branch.resblocks.7.attn.out_proj.weight is different
# key module.text_branch.resblocks.7.attn.out_proj.bias is different
# key module.text_branch.resblocks.7.ln_1.weight is different
# key module.text_branch.resblocks.7.ln_1.bias is different
# key module.text_branch.resblocks.7.mlp.c_fc.weight is different
# key module.text_branch.resblocks.7.mlp.c_fc.bias is different
# key module.text_branch.resblocks.7.mlp.c_proj.weight is different
# key module.text_branch.resblocks.7.mlp.c_proj.bias is different
# key module.text_branch.resblocks.7.ln_2.weight is different
# key module.text_branch.resblocks.7.ln_2.bias is different
# key module.text_branch.resblocks.8.attn.in_proj_weight is different
# key module.text_branch.resblocks.8.attn.in_proj_bias is different
# key module.text_branch.resblocks.8.attn.out_proj.weight is different
# key module.text_branch.resblocks.8.attn.out_proj.bias is different
# key module.text_branch.resblocks.8.ln_1.weight is different
# key module.text_branch.resblocks.8.ln_1.bias is different
# key module.text_branch.resblocks.8.mlp.c_fc.weight is different
# key module.text_branch.resblocks.8.mlp.c_fc.bias is different
# key module.text_branch.resblocks.8.mlp.c_proj.weight is different
# key module.text_branch.resblocks.8.mlp.c_proj.bias is different
# key module.text_branch.resblocks.8.ln_2.weight is different
# key module.text_branch.resblocks.8.ln_2.bias is different
# key module.text_branch.resblocks.9.attn.in_proj_weight is different
# key module.text_branch.resblocks.9.attn.in_proj_bias is different
# key module.text_branch.resblocks.9.attn.out_proj.weight is different
# key module.text_branch.resblocks.9.attn.out_proj.bias is different
# key module.text_branch.resblocks.9.ln_1.weight is different
# key module.text_branch.resblocks.9.ln_1.bias is different
# key module.text_branch.resblocks.9.mlp.c_fc.weight is different
# key module.text_branch.resblocks.9.mlp.c_fc.bias is different
# key module.text_branch.resblocks.9.mlp.c_proj.weight is different
# key module.text_branch.resblocks.9.mlp.c_proj.bias is different
# key module.text_branch.resblocks.9.ln_2.weight is different
# key module.text_branch.resblocks.9.ln_2.bias is different
# key module.text_branch.resblocks.10.attn.in_proj_weight is different
# key module.text_branch.resblocks.10.attn.in_proj_bias is different
# key module.text_branch.resblocks.10.attn.out_proj.weight is different
# key module.text_branch.resblocks.10.attn.out_proj.bias is different
# key module.text_branch.resblocks.10.ln_1.weight is different
# key module.text_branch.resblocks.10.ln_1.bias is different
# key module.text_branch.resblocks.10.mlp.c_fc.weight is different
# key module.text_branch.resblocks.10.mlp.c_fc.bias is different
# key module.text_branch.resblocks.10.mlp.c_proj.weight is different
# key module.text_branch.resblocks.10.mlp.c_proj.bias is different
# key module.text_branch.resblocks.10.ln_2.weight is different
# key module.text_branch.resblocks.10.ln_2.bias is different
# key module.text_branch.resblocks.11.attn.in_proj_weight is different
# key module.text_branch.resblocks.11.attn.in_proj_bias is different
# key module.text_branch.resblocks.11.attn.out_proj.weight is different
# key module.text_branch.resblocks.11.attn.out_proj.bias is different
# key module.text_branch.resblocks.11.ln_1.weight is different
# key module.text_branch.resblocks.11.ln_1.bias is different
# key module.text_branch.resblocks.11.mlp.c_fc.weight is different
# key module.text_branch.resblocks.11.mlp.c_fc.bias is different
# key module.text_branch.resblocks.11.mlp.c_proj.weight is different
# key module.text_branch.resblocks.11.mlp.c_proj.bias is different
# key module.text_branch.resblocks.11.ln_2.weight is different
# key module.text_branch.resblocks.11.ln_2.bias is different
# a_sum: tensor(12133.6348)
# b_sum: tensor(10423.9521)
# diff: tensor(1709.6826)
# True
# bert no freeze:
# check_ckpt_diff("/fsx/clap_logs/2022_09_14-02_33_11-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_14-02_33_11-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.encoder")
# key module.text_branch.encoder.layer.0.attention.self.query.weight is different
# key module.text_branch.encoder.layer.0.attention.self.query.bias is different
# key module.text_branch.encoder.layer.0.attention.self.key.weight is different
# key module.text_branch.encoder.layer.0.attention.self.key.bias is different
# key module.text_branch.encoder.layer.0.attention.self.value.weight is different
# key module.text_branch.encoder.layer.0.attention.self.value.bias is different
# key module.text_branch.encoder.layer.0.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.0.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.0.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.0.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.0.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.0.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.0.output.dense.weight is different
# key module.text_branch.encoder.layer.0.output.dense.bias is different
# key module.text_branch.encoder.layer.0.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.0.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.1.attention.self.query.weight is different
# key module.text_branch.encoder.layer.1.attention.self.query.bias is different
# key module.text_branch.encoder.layer.1.attention.self.key.weight is different
# key module.text_branch.encoder.layer.1.attention.self.key.bias is different
# key module.text_branch.encoder.layer.1.attention.self.value.weight is different
# key module.text_branch.encoder.layer.1.attention.self.value.bias is different
# key module.text_branch.encoder.layer.1.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.1.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.1.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.1.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.1.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.1.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.1.output.dense.weight is different
# key module.text_branch.encoder.layer.1.output.dense.bias is different
# key module.text_branch.encoder.layer.1.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.1.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.2.attention.self.query.weight is different
# key module.text_branch.encoder.layer.2.attention.self.query.bias is different
# key module.text_branch.encoder.layer.2.attention.self.key.weight is different
# key module.text_branch.encoder.layer.2.attention.self.key.bias is different
# key module.text_branch.encoder.layer.2.attention.self.value.weight is different
# key module.text_branch.encoder.layer.2.attention.self.value.bias is different
# key module.text_branch.encoder.layer.2.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.2.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.2.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.2.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.2.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.2.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.2.output.dense.weight is different
# key module.text_branch.encoder.layer.2.output.dense.bias is different
# key module.text_branch.encoder.layer.2.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.2.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.3.attention.self.query.weight is different
# key module.text_branch.encoder.layer.3.attention.self.query.bias is different
# key module.text_branch.encoder.layer.3.attention.self.key.weight is different
# key module.text_branch.encoder.layer.3.attention.self.key.bias is different
# key module.text_branch.encoder.layer.3.attention.self.value.weight is different
# key module.text_branch.encoder.layer.3.attention.self.value.bias is different
# key module.text_branch.encoder.layer.3.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.3.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.3.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.3.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.3.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.3.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.3.output.dense.weight is different
# key module.text_branch.encoder.layer.3.output.dense.bias is different
# key module.text_branch.encoder.layer.3.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.3.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.4.attention.self.query.weight is different
# key module.text_branch.encoder.layer.4.attention.self.query.bias is different
# key module.text_branch.encoder.layer.4.attention.self.key.weight is different
# key module.text_branch.encoder.layer.4.attention.self.key.bias is different
# key module.text_branch.encoder.layer.4.attention.self.value.weight is different
# key module.text_branch.encoder.layer.4.attention.self.value.bias is different
# key module.text_branch.encoder.layer.4.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.4.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.4.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.4.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.4.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.4.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.4.output.dense.weight is different
# key module.text_branch.encoder.layer.4.output.dense.bias is different
# key module.text_branch.encoder.layer.4.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.4.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.5.attention.self.query.weight is different
# key module.text_branch.encoder.layer.5.attention.self.query.bias is different
# key module.text_branch.encoder.layer.5.attention.self.key.weight is different
# key module.text_branch.encoder.layer.5.attention.self.key.bias is different
# key module.text_branch.encoder.layer.5.attention.self.value.weight is different
# key module.text_branch.encoder.layer.5.attention.self.value.bias is different
# key module.text_branch.encoder.layer.5.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.5.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.5.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.5.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.5.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.5.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.5.output.dense.weight is different
# key module.text_branch.encoder.layer.5.output.dense.bias is different
# key module.text_branch.encoder.layer.5.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.5.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.6.attention.self.query.weight is different
# key module.text_branch.encoder.layer.6.attention.self.query.bias is different
# key module.text_branch.encoder.layer.6.attention.self.key.weight is different
# key module.text_branch.encoder.layer.6.attention.self.key.bias is different
# key module.text_branch.encoder.layer.6.attention.self.value.weight is different
# key module.text_branch.encoder.layer.6.attention.self.value.bias is different
# key module.text_branch.encoder.layer.6.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.6.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.6.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.6.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.6.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.6.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.6.output.dense.weight is different
# key module.text_branch.encoder.layer.6.output.dense.bias is different
# key module.text_branch.encoder.layer.6.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.6.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.7.attention.self.query.weight is different
# key module.text_branch.encoder.layer.7.attention.self.query.bias is different
# key module.text_branch.encoder.layer.7.attention.self.key.weight is different
# key module.text_branch.encoder.layer.7.attention.self.key.bias is different
# key module.text_branch.encoder.layer.7.attention.self.value.weight is different
# key module.text_branch.encoder.layer.7.attention.self.value.bias is different
# key module.text_branch.encoder.layer.7.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.7.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.7.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.7.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.7.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.7.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.7.output.dense.weight is different
# key module.text_branch.encoder.layer.7.output.dense.bias is different
# key module.text_branch.encoder.layer.7.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.7.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.8.attention.self.query.weight is different
# key module.text_branch.encoder.layer.8.attention.self.query.bias is different
# key module.text_branch.encoder.layer.8.attention.self.key.weight is different
# key module.text_branch.encoder.layer.8.attention.self.key.bias is different
# key module.text_branch.encoder.layer.8.attention.self.value.weight is different
# key module.text_branch.encoder.layer.8.attention.self.value.bias is different
# key module.text_branch.encoder.layer.8.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.8.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.8.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.8.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.8.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.8.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.8.output.dense.weight is different
# key module.text_branch.encoder.layer.8.output.dense.bias is different
# key module.text_branch.encoder.layer.8.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.8.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.9.attention.self.query.weight is different
# key module.text_branch.encoder.layer.9.attention.self.query.bias is different
# key module.text_branch.encoder.layer.9.attention.self.key.weight is different
# key module.text_branch.encoder.layer.9.attention.self.key.bias is different
# key module.text_branch.encoder.layer.9.attention.self.value.weight is different
# key module.text_branch.encoder.layer.9.attention.self.value.bias is different
# key module.text_branch.encoder.layer.9.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.9.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.9.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.9.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.9.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.9.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.9.output.dense.weight is different
# key module.text_branch.encoder.layer.9.output.dense.bias is different
# key module.text_branch.encoder.layer.9.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.9.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.10.attention.self.query.weight is different
# key module.text_branch.encoder.layer.10.attention.self.query.bias is different
# key module.text_branch.encoder.layer.10.attention.self.key.weight is different
# key module.text_branch.encoder.layer.10.attention.self.key.bias is different
# key module.text_branch.encoder.layer.10.attention.self.value.weight is different
# key module.text_branch.encoder.layer.10.attention.self.value.bias is different
# key module.text_branch.encoder.layer.10.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.10.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.10.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.10.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.10.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.10.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.10.output.dense.weight is different
# key module.text_branch.encoder.layer.10.output.dense.bias is different
# key module.text_branch.encoder.layer.10.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.10.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.11.attention.self.query.weight is different
# key module.text_branch.encoder.layer.11.attention.self.query.bias is different
# key module.text_branch.encoder.layer.11.attention.self.key.weight is different
# key module.text_branch.encoder.layer.11.attention.self.key.bias is different
# key module.text_branch.encoder.layer.11.attention.self.value.weight is different
# key module.text_branch.encoder.layer.11.attention.self.value.bias is different
# key module.text_branch.encoder.layer.11.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.11.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.11.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.11.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.11.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.11.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.11.output.dense.weight is different
# key module.text_branch.encoder.layer.11.output.dense.bias is different
# key module.text_branch.encoder.layer.11.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.11.output.LayerNorm.bias is different
# a_sum: tensor(15185.1230)
# b_sum: tensor(15576.5596)
# diff: tensor(-391.4365)
# True
# bert freeze:
# check_ckpt_diff("/fsx/clap_logs/2022_09_13-01_25_15-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_10.pt", "/fsx/clap_logs/2022_09_13-01_25_15-model_PANN-14-lr_0.0001-b_160-j_4-p_fp32/checkpoints/epoch_100.pt", "text_branch.encoder")
# key module.text_branch.encoder.layer.0.attention.self.query.weight is different
# key module.text_branch.encoder.layer.0.attention.self.query.bias is different
# key module.text_branch.encoder.layer.0.attention.self.key.weight is different
# key module.text_branch.encoder.layer.0.attention.self.key.bias is different
# key module.text_branch.encoder.layer.0.attention.self.value.weight is different
# key module.text_branch.encoder.layer.0.attention.self.value.bias is different
# key module.text_branch.encoder.layer.0.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.0.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.0.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.0.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.0.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.0.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.0.output.dense.weight is different
# key module.text_branch.encoder.layer.0.output.dense.bias is different
# key module.text_branch.encoder.layer.0.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.0.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.1.attention.self.query.weight is different
# key module.text_branch.encoder.layer.1.attention.self.query.bias is different
# key module.text_branch.encoder.layer.1.attention.self.key.weight is different
# key module.text_branch.encoder.layer.1.attention.self.key.bias is different
# key module.text_branch.encoder.layer.1.attention.self.value.weight is different
# key module.text_branch.encoder.layer.1.attention.self.value.bias is different
# key module.text_branch.encoder.layer.1.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.1.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.1.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.1.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.1.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.1.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.1.output.dense.weight is different
# key module.text_branch.encoder.layer.1.output.dense.bias is different
# key module.text_branch.encoder.layer.1.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.1.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.2.attention.self.query.weight is different
# key module.text_branch.encoder.layer.2.attention.self.query.bias is different
# key module.text_branch.encoder.layer.2.attention.self.key.weight is different
# key module.text_branch.encoder.layer.2.attention.self.key.bias is different
# key module.text_branch.encoder.layer.2.attention.self.value.weight is different
# key module.text_branch.encoder.layer.2.attention.self.value.bias is different
# key module.text_branch.encoder.layer.2.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.2.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.2.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.2.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.2.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.2.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.2.output.dense.weight is different
# key module.text_branch.encoder.layer.2.output.dense.bias is different
# key module.text_branch.encoder.layer.2.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.2.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.3.attention.self.query.weight is different
# key module.text_branch.encoder.layer.3.attention.self.query.bias is different
# key module.text_branch.encoder.layer.3.attention.self.key.weight is different
# key module.text_branch.encoder.layer.3.attention.self.key.bias is different
# key module.text_branch.encoder.layer.3.attention.self.value.weight is different
# key module.text_branch.encoder.layer.3.attention.self.value.bias is different
# key module.text_branch.encoder.layer.3.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.3.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.3.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.3.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.3.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.3.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.3.output.dense.weight is different
# key module.text_branch.encoder.layer.3.output.dense.bias is different
# key module.text_branch.encoder.layer.3.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.3.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.4.attention.self.query.weight is different
# key module.text_branch.encoder.layer.4.attention.self.query.bias is different
# key module.text_branch.encoder.layer.4.attention.self.key.weight is different
# key module.text_branch.encoder.layer.4.attention.self.key.bias is different
# key module.text_branch.encoder.layer.4.attention.self.value.weight is different
# key module.text_branch.encoder.layer.4.attention.self.value.bias is different
# key module.text_branch.encoder.layer.4.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.4.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.4.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.4.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.4.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.4.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.4.output.dense.weight is different
# key module.text_branch.encoder.layer.4.output.dense.bias is different
# key module.text_branch.encoder.layer.4.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.4.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.5.attention.self.query.weight is different
# key module.text_branch.encoder.layer.5.attention.self.query.bias is different
# key module.text_branch.encoder.layer.5.attention.self.key.weight is different
# key module.text_branch.encoder.layer.5.attention.self.key.bias is different
# key module.text_branch.encoder.layer.5.attention.self.value.weight is different
# key module.text_branch.encoder.layer.5.attention.self.value.bias is different
# key module.text_branch.encoder.layer.5.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.5.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.5.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.5.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.5.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.5.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.5.output.dense.weight is different
# key module.text_branch.encoder.layer.5.output.dense.bias is different
# key module.text_branch.encoder.layer.5.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.5.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.6.attention.self.query.weight is different
# key module.text_branch.encoder.layer.6.attention.self.query.bias is different
# key module.text_branch.encoder.layer.6.attention.self.key.weight is different
# key module.text_branch.encoder.layer.6.attention.self.key.bias is different
# key module.text_branch.encoder.layer.6.attention.self.value.weight is different
# key module.text_branch.encoder.layer.6.attention.self.value.bias is different
# key module.text_branch.encoder.layer.6.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.6.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.6.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.6.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.6.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.6.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.6.output.dense.weight is different
# key module.text_branch.encoder.layer.6.output.dense.bias is different
# key module.text_branch.encoder.layer.6.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.6.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.7.attention.self.query.weight is different
# key module.text_branch.encoder.layer.7.attention.self.query.bias is different
# key module.text_branch.encoder.layer.7.attention.self.key.weight is different
# key module.text_branch.encoder.layer.7.attention.self.key.bias is different
# key module.text_branch.encoder.layer.7.attention.self.value.weight is different
# key module.text_branch.encoder.layer.7.attention.self.value.bias is different
# key module.text_branch.encoder.layer.7.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.7.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.7.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.7.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.7.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.7.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.7.output.dense.weight is different
# key module.text_branch.encoder.layer.7.output.dense.bias is different
# key module.text_branch.encoder.layer.7.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.7.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.8.attention.self.query.weight is different
# key module.text_branch.encoder.layer.8.attention.self.query.bias is different
# key module.text_branch.encoder.layer.8.attention.self.key.weight is different
# key module.text_branch.encoder.layer.8.attention.self.key.bias is different
# key module.text_branch.encoder.layer.8.attention.self.value.weight is different
# key module.text_branch.encoder.layer.8.attention.self.value.bias is different
# key module.text_branch.encoder.layer.8.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.8.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.8.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.8.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.8.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.8.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.8.output.dense.weight is different
# key module.text_branch.encoder.layer.8.output.dense.bias is different
# key module.text_branch.encoder.layer.8.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.8.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.9.attention.self.query.weight is different
# key module.text_branch.encoder.layer.9.attention.self.query.bias is different
# key module.text_branch.encoder.layer.9.attention.self.key.weight is different
# key module.text_branch.encoder.layer.9.attention.self.key.bias is different
# key module.text_branch.encoder.layer.9.attention.self.value.weight is different
# key module.text_branch.encoder.layer.9.attention.self.value.bias is different
# key module.text_branch.encoder.layer.9.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.9.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.9.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.9.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.9.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.9.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.9.output.dense.weight is different
# key module.text_branch.encoder.layer.9.output.dense.bias is different
# key module.text_branch.encoder.layer.9.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.9.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.10.attention.self.query.weight is different
# key module.text_branch.encoder.layer.10.attention.self.query.bias is different
# key module.text_branch.encoder.layer.10.attention.self.key.weight is different
# key module.text_branch.encoder.layer.10.attention.self.key.bias is different
# key module.text_branch.encoder.layer.10.attention.self.value.weight is different
# key module.text_branch.encoder.layer.10.attention.self.value.bias is different
# key module.text_branch.encoder.layer.10.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.10.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.10.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.10.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.10.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.10.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.10.output.dense.weight is different
# key module.text_branch.encoder.layer.10.output.dense.bias is different
# key module.text_branch.encoder.layer.10.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.10.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.11.attention.self.query.weight is different
# key module.text_branch.encoder.layer.11.attention.self.query.bias is different
# key module.text_branch.encoder.layer.11.attention.self.key.weight is different
# key module.text_branch.encoder.layer.11.attention.self.key.bias is different
# key module.text_branch.encoder.layer.11.attention.self.value.weight is different
# key module.text_branch.encoder.layer.11.attention.self.value.bias is different
# key module.text_branch.encoder.layer.11.attention.output.dense.weight is different
# key module.text_branch.encoder.layer.11.attention.output.dense.bias is different
# key module.text_branch.encoder.layer.11.attention.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.11.attention.output.LayerNorm.bias is different
# key module.text_branch.encoder.layer.11.intermediate.dense.weight is different
# key module.text_branch.encoder.layer.11.intermediate.dense.bias is different
# key module.text_branch.encoder.layer.11.output.dense.weight is different
# key module.text_branch.encoder.layer.11.output.dense.bias is different
# key module.text_branch.encoder.layer.11.output.LayerNorm.weight is different
# key module.text_branch.encoder.layer.11.output.LayerNorm.bias is different
# a_sum: tensor(15078.6641)
# b_sum: tensor(15540.0723)
# diff: tensor(-461.4082)
# True
# linear_prob_text
# check_ckpt_diff("/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_50.pt", "/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_100.pt", "text_branch.resblocks")
# a_sum: tensor(12111.0244)
# b_sum: tensor(12111.0244)
# diff: tensor(0.)
# linear_prob_audio
# check_ckpt_diff("/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_50.pt", "/fsx/clap_logs/2022_09_15-02_05_29-linear_probemodel_PANN-14-lr_0.0001-b_512-j_4-p_fp32/checkpoints/pretrain_epoch_10_lp_epoch_100.pt", "clap_model")
# key clap_model.audio_branch.bn0.num_batches_tracked is different
# key clap_model.audio_branch.conv_block1.bn1.running_mean is different
# key clap_model.audio_branch.conv_block1.bn1.running_var is different
# key clap_model.audio_branch.conv_block1.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block1.bn2.running_mean is different
# key clap_model.audio_branch.conv_block1.bn2.running_var is different
# key clap_model.audio_branch.conv_block1.bn2.num_batches_tracked is different
# key clap_model.audio_branch.conv_block2.bn1.running_mean is different
# key clap_model.audio_branch.conv_block2.bn1.running_var is different
# key clap_model.audio_branch.conv_block2.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block2.bn2.running_mean is different
# key clap_model.audio_branch.conv_block2.bn2.running_var is different
# key clap_model.audio_branch.conv_block2.bn2.num_batches_tracked is different
# key clap_model.audio_branch.conv_block3.bn1.running_mean is different
# key clap_model.audio_branch.conv_block3.bn1.running_var is different
# key clap_model.audio_branch.conv_block3.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block3.bn2.running_mean is different
# key clap_model.audio_branch.conv_block3.bn2.running_var is different
# key clap_model.audio_branch.conv_block3.bn2.num_batches_tracked is different
# key clap_model.audio_branch.conv_block4.bn1.running_mean is different
# key clap_model.audio_branch.conv_block4.bn1.running_var is different
# key clap_model.audio_branch.conv_block4.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block4.bn2.running_mean is different
# key clap_model.audio_branch.conv_block4.bn2.running_var is different
# key clap_model.audio_branch.conv_block4.bn2.num_batches_tracked is different
# key clap_model.audio_branch.conv_block5.bn1.running_mean is different
# key clap_model.audio_branch.conv_block5.bn1.running_var is different
# key clap_model.audio_branch.conv_block5.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block5.bn2.running_mean is different
# key clap_model.audio_branch.conv_block5.bn2.running_var is different
# key clap_model.audio_branch.conv_block5.bn2.num_batches_tracked is different
# key clap_model.audio_branch.conv_block6.bn1.running_mean is different
# key clap_model.audio_branch.conv_block6.bn1.running_var is different
# key clap_model.audio_branch.conv_block6.bn1.num_batches_tracked is different
# key clap_model.audio_branch.conv_block6.bn2.running_mean is different
# key clap_model.audio_branch.conv_block6.bn2.running_var is different
# key clap_model.audio_branch.conv_block6.bn2.num_batches_tracked is different
# a_sum: tensor(120061.5078)
# b_sum: tensor(122656.0469)
# diff: tensor(-2594.5391)
# True