mjtwins commited on
Commit
7ef898a
1 Parent(s): b392c7b

add sparse model

Browse files
Files changed (4) hide show
  1. config.json +7 -11
  2. pytorch_model.bin +2 -2
  3. sparse_encoder.pk +0 -0
  4. sparse_weight.pt +3 -0
config.json CHANGED
@@ -1,10 +1,6 @@
1
  {
2
- "_name_or_path": "dmis-lab/biobert-base-cased-v1.1",
3
- "architectures": [
4
- "BertModel"
5
- ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
  "hidden_act": "gelu",
9
  "hidden_dropout_prob": 0.1,
10
  "hidden_size": 768,
@@ -12,14 +8,14 @@
12
  "intermediate_size": 3072,
13
  "layer_norm_eps": 1e-12,
14
  "max_position_embeddings": 512,
15
- "model_type": "bert",
16
  "num_attention_heads": 12,
17
  "num_hidden_layers": 12,
18
- "pad_token_id": 0,
19
- "position_embedding_type": "absolute",
20
- "torch_dtype": "float32",
21
- "transformers_version": "4.11.3",
 
22
  "type_vocab_size": 2,
23
- "use_cache": true,
24
  "vocab_size": 28996
25
  }
1
  {
 
 
 
 
2
  "attention_probs_dropout_prob": 0.1,
3
+ "finetuning_task": null,
4
  "hidden_act": "gelu",
5
  "hidden_dropout_prob": 0.1,
6
  "hidden_size": 768,
8
  "intermediate_size": 3072,
9
  "layer_norm_eps": 1e-12,
10
  "max_position_embeddings": 512,
 
11
  "num_attention_heads": 12,
12
  "num_hidden_layers": 12,
13
+ "num_labels": 2,
14
+ "output_attentions": false,
15
+ "output_hidden_states": false,
16
+ "pruned_heads": {},
17
+ "torchscript": false,
18
  "type_vocab_size": 2,
19
+ "use_bfloat16": false,
20
  "vocab_size": 28996
21
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a5a38731e458a8d8148817b17721e99425954225c570a9de52c3256568e21413
3
- size 433327607
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5b64d4ec878852e0bdd02ed69d3ad9d9371697ac90685740d7ef1aeddd81fc1
3
+ size 433288887
sparse_encoder.pk ADDED
Binary file (66.9 kB). View file
sparse_weight.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52a155effbbc735d83a5871b7ca7f20307eb61217d2ac24d1734f880ad568205
3
+ size 392