Your Name
new app
d5cd13d
optimus:
symbol: optimus
find_unused_parameters: false
args: {}
optimus_bert_encoder:
super_cfg: optimus
type: optimus_bert_connector
# pth: pretrained/optimus_bert_encoder.pth
args:
config:
architectures:
- BertForMaskedLM
attention_probs_dropout_prob: 0.1
finetuning_task: null
hidden_act: gelu
hidden_dropout_prob: 0.1
hidden_size: 768
initializer_range: 0.02
intermediate_size: 3072
layer_norm_eps: 1.e-12
max_position_embeddings: 512
num_attention_heads: 12
num_hidden_layers: 12
num_labels: 2
output_attentions: false
output_hidden_states: false
pruned_heads: {}
torchscript: false
type_vocab_size: 2
vocab_size: 28996
latent_size: 768
optimus_bert_tokenizer:
super_cfg: optimus
type: optimus_bert_tokenizer
args:
do_lower_case: false
max_len: 512
vocab_file: lib/model_zoo/optimus_models/vocab/bert-base-cased-vocab.txt
optimus_gpt2_decoder:
super_cfg: optimus
type: optimus_gpt2_connector
# pth: pretrained/optimus_gpt2_decoder.pth
args:
config:
architectures:
- GPT2LMHeadModel
attn_pdrop: 0.1
embd_pdrop: 0.1
finetuning_task: null
hidden_size: 768
initializer_range: 0.02
latent_size: 768
layer_norm_epsilon: 1.e-05
max_position_embeddings: 1024
n_ctx: 1024
n_embd: 768
n_head: 12
n_layer: 12
n_positions: 1024
num_attention_heads: 12
num_hidden_layers: 12
num_labels: 1
output_attentions: false
output_hidden_states: false
pretrained_config_archive_map:
gpt2 : https://s3.amazonaws.com/models.huggingface.co/bert/gpt2-config.json
gpt2-medium : https://s3.amazonaws.com/models.huggingface.co/bert/gpt2-medium-config.json
gpt2-large : https://s3.amazonaws.com/models.huggingface.co/bert/gpt2-large-config.json
pruned_heads: {}
resid_pdrop: 0.1
summary_activation: null
summary_first_dropout: 0.1
summary_proj_to_labels: true
summary_type: cls_index
summary_use_proj: true
torchscript: false
vocab_size: 50260
optimus_gpt2_tokenizer:
super_cfg: optimus
type: optimus_gpt2_tokenizer
args:
do_lower_case: false
max_len: 1024
vocab_file: lib/model_zoo/optimus_models/vocab/gpt2-vocab.json
merges_file: lib/model_zoo/optimus_models/vocab/gpt2-merges.txt
optimus_v1:
super_cfg: optimus
type: optimus_vae_next
args:
encoder: MODEL(optimus_bert_encoder)
decoder: MODEL(optimus_gpt2_decoder)
tokenizer_encoder: MODEL(optimus_bert_tokenizer)
tokenizer_decoder: MODEL(optimus_gpt2_tokenizer)
args:
latent_size: 768
# pth: pretrained/optimus-vae.pth
hfm: ['shi-labs/versatile-diffusion-model', 'pretrained_pth/optimus-vae.pth']