File size: 961 Bytes
bc1f1f4
 
 
5b5da1b
7508c02
13649aa
ca83fd7
5b5da1b
7508c02
ca83fd7
 
bc1f1f4
 
 
ebf2390
 
 
 
 
 
bc1f1f4
ebf2390
bc1f1f4
ca83fd7
bc1f1f4
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
from omegaconf import OmegaConf
from scripts.rendertext_tool import Render_Text, load_model_from_config
import torch

# cfg = OmegaConf.load("other_configs/config_ema.yaml")
# model = load_model_from_config(cfg, "model_states.pt", verbose=True)
# model = load_model_from_config(cfg, "mp_rank_00_model_states.pt", verbose=True)

cfg = OmegaConf.load("other_configs/config_ema_unlock.yaml")
epoch_idx = 39
model = load_model_from_config(cfg, "epoch={:0>6d}.ckpt".format(epoch_idx), verbose=True)

from pytorch_lightning.callbacks import ModelCheckpoint
with model.ema_scope("store ema weights"):
    model_sd = model.state_dict()
    store_sd = {}
    for key in model_sd:
        if "ema" in key:
            continue
        store_sd[key] = model_sd[key]
    file_content = {
        'state_dict': store_sd
    }
    torch.save(file_content, f"textcaps5K_epoch_{epoch_idx+1}_model_wo_ema.ckpt")
    print("has stored the transfered ckpt.")
print("trial ends!")