philschmid HF staff commited on
Commit
6ffff56
1 Parent(s): 33e4cde

model version

Browse files
.DS_Store ADDED
Binary file (6.15 kB). View file
 
all_results.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.9013761467889908,
4
+ "eval_loss": 0.3691919445991516,
5
+ "eval_mem_cpu_alloc_delta": 1605632,
6
+ "eval_mem_cpu_peaked_delta": 0,
7
+ "eval_mem_gpu_alloc_delta": 0,
8
+ "eval_mem_gpu_peaked_delta": 25477120,
9
+ "eval_runtime": 3.9152,
10
+ "eval_samples": 872,
11
+ "eval_samples_per_second": 222.719,
12
+ "init_mem_cpu_alloc_delta": 1199738880,
13
+ "init_mem_cpu_peaked_delta": 74649600,
14
+ "init_mem_gpu_alloc_delta": 90860544,
15
+ "init_mem_gpu_peaked_delta": 0,
16
+ "train_mem_cpu_alloc_delta": 434720768,
17
+ "train_mem_cpu_peaked_delta": 141451264,
18
+ "train_mem_gpu_alloc_delta": 274994176,
19
+ "train_mem_gpu_peaked_delta": 1159366656,
20
+ "train_runtime": 1085.2317,
21
+ "train_samples": 67349,
22
+ "train_samples_per_second": 5.819
23
+ }
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "nreimers/MiniLM-L6-H384-uncased",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "finetuning_task": "sst2",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 384,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 1536,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 6,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "problem_type": "single_label_classification",
22
+ "transformers_version": "4.6.1",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
eval_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.9013761467889908,
4
+ "eval_loss": 0.3691919445991516,
5
+ "eval_mem_cpu_alloc_delta": 1605632,
6
+ "eval_mem_cpu_peaked_delta": 0,
7
+ "eval_mem_gpu_alloc_delta": 0,
8
+ "eval_mem_gpu_peaked_delta": 25477120,
9
+ "eval_runtime": 3.9152,
10
+ "eval_samples": 872,
11
+ "eval_samples_per_second": 222.719
12
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51107694bb6266b86e5951871995895b1c95438fc35b292e4c06e113196018f7
3
+ size 90904553
special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "name_or_path": "nreimers/MiniLM-L6-H384-uncased", "do_basic_tokenize": true, "never_split": null, "model_max_length": 512, "special_tokens_map_file": "/root/.cache/huggingface/transformers/4bab06d0845c01b218fc031866e62ccf55a15d3f649549f76c89e002ba24d5d9.dd8bd9bfd3664b530ea4e645105f557769387b3da9f79bdb55ed556bdd80611d"}
train_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "init_mem_cpu_alloc_delta": 1199738880,
4
+ "init_mem_cpu_peaked_delta": 74649600,
5
+ "init_mem_gpu_alloc_delta": 90860544,
6
+ "init_mem_gpu_peaked_delta": 0,
7
+ "train_mem_cpu_alloc_delta": 434720768,
8
+ "train_mem_cpu_peaked_delta": 141451264,
9
+ "train_mem_gpu_alloc_delta": 274994176,
10
+ "train_mem_gpu_peaked_delta": 1159366656,
11
+ "train_runtime": 1085.2317,
12
+ "train_samples": 67349,
13
+ "train_samples_per_second": 5.819
14
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff