mgh6 commited on
Commit
d686dd2
·
1 Parent(s): e9c1e6f

Training in progress, step 500

Browse files
config.json CHANGED
@@ -1,27 +1,27 @@
1
  {
2
- "_name_or_path": "facebook/esm1b_t33_650M_UR50S",
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
8
- "emb_layer_norm_before": true,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.0,
12
- "hidden_size": 1280,
13
  "initializer_range": 0.02,
14
- "intermediate_size": 5120,
15
  "is_folding_model": false,
16
  "layer_norm_eps": 1e-05,
17
  "mask_token_id": 32,
18
  "max_position_embeddings": 1026,
19
  "model_type": "esm",
20
  "num_attention_heads": 20,
21
- "num_hidden_layers": 33,
22
  "output_attentions": true,
23
  "pad_token_id": 1,
24
- "position_embedding_type": "absolute",
25
  "token_dropout": true,
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.31.0",
 
1
  {
2
+ "_name_or_path": "facebook/esm2_t12_35M_UR50D",
3
  "architectures": [
4
  "EsmForMaskedLM"
5
  ],
6
  "attention_probs_dropout_prob": 0.0,
7
  "classifier_dropout": null,
8
+ "emb_layer_norm_before": false,
9
  "esmfold_config": null,
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.0,
12
+ "hidden_size": 480,
13
  "initializer_range": 0.02,
14
+ "intermediate_size": 1920,
15
  "is_folding_model": false,
16
  "layer_norm_eps": 1e-05,
17
  "mask_token_id": 32,
18
  "max_position_embeddings": 1026,
19
  "model_type": "esm",
20
  "num_attention_heads": 20,
21
+ "num_hidden_layers": 12,
22
  "output_attentions": true,
23
  "pad_token_id": 1,
24
+ "position_embedding_type": "rotary",
25
  "token_dropout": true,
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.31.0",
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:668e3f23175e955a4f9771f149797dd14fbf6e6697f253480be1d409f2f1747d
3
- size 2609620225
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b607742916a1ba75e48a0b40df597bc784a1dbf757abf4e975ca0567fccee338
3
+ size 136048297
runs/Jul23_21-05-29_4ab41e0c8300/events.out.tfevents.1690146335.4ab41e0c8300.32683.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b8507e4d7b42c9f946040b763a928756c5de9c83393113b82287f89dfe57032
3
+ size 4161
runs/Jul23_21-05-58_4ab41e0c8300/events.out.tfevents.1690146361.4ab41e0c8300.32683.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c5089fcdd4a5a1f9c420cdb8f459840666139c598ce202ad4e98e6112e40169d
3
+ size 4862
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f6df29bb7664bd68250f0bdb970548df95c9acffd0ed06ecddd05295347a2dcf
3
  size 3963
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3eb365d361fc639e289da00527dca75cf7387b45dec7744e55d9e33616e298c
3
  size 3963