lewtun HF staff commited on
Commit
de61cb9
1 Parent(s): 3107a94
.gitignore ADDED
@@ -0,0 +1 @@
 
1
+ checkpoint-*/
README.md ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - generated_from_trainer
5
+ datasets:
6
+ - imdb
7
+ model-index:
8
+ - name: bert-base-uncased-finetuned-imdb
9
+ results:
10
+ - task:
11
+ name: Masked Language Modeling
12
+ type: fill-mask
13
+ dataset:
14
+ name: imdb
15
+ type: imdb
16
+ args: plain_text
17
+ ---
18
+
19
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
20
+ should probably proofread and complete it, then remove this comment. -->
21
+
22
+ # bert-base-uncased-finetuned-imdb
23
+
24
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the imdb dataset.
25
+ It achieves the following results on the evaluation set:
26
+ - Loss: 2.0284
27
+
28
+ ## Model description
29
+
30
+ More information needed
31
+
32
+ ## Intended uses & limitations
33
+
34
+ More information needed
35
+
36
+ ## Training and evaluation data
37
+
38
+ More information needed
39
+
40
+ ## Training procedure
41
+
42
+ ### Training hyperparameters
43
+
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 2e-05
46
+ - train_batch_size: 64
47
+ - eval_batch_size: 64
48
+ - seed: 42
49
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
+ - lr_scheduler_type: linear
51
+ - num_epochs: 3.0
52
+ - mixed_precision_training: Native AMP
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss |
57
+ |:-------------:|:-----:|:----:|:---------------:|
58
+ | 2.2244 | 1.0 | 958 | 2.0726 |
59
+ | 2.1537 | 2.0 | 1916 | 2.0381 |
60
+ | 2.1183 | 3.0 | 2874 | 2.0284 |
61
+
62
+
63
+ ### Framework versions
64
+
65
+ - Transformers 4.10.3
66
+ - Pytorch 1.9.1+cu111
67
+ - Datasets 1.12.1
68
+ - Tokenizers 0.10.3
config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "bert-base-uncased",
3
+ "architectures": [
4
+ "BertForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 768,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 3072,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 12,
18
+ "num_hidden_layers": 12,
19
+ "pad_token_id": 0,
20
+ "position_embedding_type": "absolute",
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.10.3",
23
+ "type_vocab_size": 2,
24
+ "use_cache": true,
25
+ "vocab_size": 30522
26
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb4d7370fafafc61d0ef312503f620c23208fd604d3db0d38946937ccb35bfcc
3
+ size 438141995
runs/Sep28_22-24-24_vorace/1632860714.3525183/events.out.tfevents.1632860714.vorace.2629792.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b628eb7ccce50a587af7e5082af864a5e0fb6e648da8682d351b72d094bc6fe6
3
+ size 4232
runs/Sep28_22-24-24_vorace/events.out.tfevents.1632860714.vorace.2629792.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:161254a542f940ad7c92f61ef23ff2e4f6a9ede36782cc0cde08720935723068
3
+ size 5261
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65c8ff14d2335ed3ac702edaf15cb704d7380d9e9d4344bea9984eff2c7d3bc7
3
+ size 2671