yuxindu's picture
init the space
9850295
raw
history blame
No virus
3.55 kB
# Copyright (c) Meta Platforms, Inc. and affiliates.
# All rights reserved.
# This source code is licensed under the license found in the
# LICENSE file in the root directory of this source tree.
from functools import partial
from pathlib import Path
import urllib.request
import torch
from .modeling import (
ImageEncoderViT,
MaskDecoder,
PromptEncoder,
Sam,
TwoWayTransformer,
)
import numpy as np
from .modeling.image_encoder_swin import SwinTransformer
from monai.networks.nets import ViT
from monai.networks.nets.swin_unetr import SwinTransformer as SwinViT
from monai.utils import ensure_tuple_rep, optional_import
"""
Examples::
# for 3D single channel input with size (96,96,96), 4-channel output and feature size of 48.
>>> net = SwinUNETR(img_size=(96,96,96), in_channels=1, out_channels=4, feature_size=48)
# for 3D 4-channel input with size (128,128,128), 3-channel output and (2,4,2,2) layers in each stage.
>>> net = SwinUNETR(img_size=(128,128,128), in_channels=4, out_channels=3, depths=(2,4,2,2))
# for 2D single channel input with size (96,96), 2-channel output and gradient checkpointing.
>>> net = SwinUNETR(img_size=(96,96), in_channels=3, out_channels=2, use_checkpoint=True, spatial_dims=2)
"""
def build_sam_vit_3d(checkpoint=None):
print('build_sam_vit_3d...')
return _build_sam(
image_encoder_type='vit',
embed_dim = 768,
patch_size=[4,16,16],
checkpoint=checkpoint,
image_size=[32,256,256],
)
sam_model_registry = {
"vit": build_sam_vit_3d,
}
def _build_sam(
image_encoder_type,
embed_dim,
patch_size,
checkpoint,
image_size,
):
mlp_dim = 3072
num_layers = 12
num_heads = 12
pos_embed = 'perceptron'
dropout_rate = 0.0
image_encoder=ViT(
in_channels=1,
img_size=image_size,
patch_size=patch_size,
hidden_size=embed_dim,
mlp_dim=mlp_dim,
num_layers=num_layers,
num_heads=num_heads,
pos_embed=pos_embed,
classification=False,
dropout_rate=dropout_rate,
)
image_embedding_size = [int(item) for item in (np.array(image_size) / np.array(patch_size))]
if checkpoint is not None:
with open(checkpoint, "rb") as f:
state_dict = torch.load(f, map_location='cpu')['state_dict']
encoder_dict = {k.replace('model.encoder.', ''): v for k, v in state_dict.items() if 'model.encoder.' in k}
image_encoder.load_state_dict(encoder_dict)
print(f'===> image_encoder.load_param: {checkpoint}')
sam = Sam(
image_encoder=image_encoder,
prompt_encoder=PromptEncoder(
embed_dim=embed_dim,
image_embedding_size=image_embedding_size,
input_image_size=image_size,
mask_in_chans=16,
),
mask_decoder=MaskDecoder(
image_encoder_type=image_encoder_type,
num_multimask_outputs=3,
transformer=TwoWayTransformer(
depth=2,
embedding_dim=embed_dim,
mlp_dim=2048,
num_heads=8,
),
transformer_dim=embed_dim,
iou_head_depth=3,
iou_head_hidden_dim=256,
image_size=np.array(image_size),
patch_size=np.array(patch_size),
),
pixel_mean=[123.675, 116.28, 103.53],
pixel_std=[58.395, 57.12, 57.375],
)
sam.eval()
return sam