willherbert27 commited on
Commit
05221fd
1 Parent(s): d27a1d2

willherbert27/bert-finetuned-no-context-squad

Browse files
Files changed (4) hide show
  1. README.md +66 -0
  2. model.safetensors +1 -1
  3. tokenizer.json +2 -16
  4. training_args.bin +1 -1
README.md ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: willherbert27/bert-finetuned-combo-textbook-no-context
4
+ tags:
5
+ - generated_from_trainer
6
+ model-index:
7
+ - name: bert-textbook-no-context-finetuned-squad
8
+ results: []
9
+ ---
10
+
11
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
+ should probably proofread and complete it, then remove this comment. -->
13
+
14
+ # bert-textbook-no-context-finetuned-squad
15
+
16
+ This model is a fine-tuned version of [willherbert27/bert-finetuned-combo-textbook-no-context](https://huggingface.co/willherbert27/bert-finetuned-combo-textbook-no-context) on an unknown dataset.
17
+ It achieves the following results on the evaluation set:
18
+ - Loss: 4.2753
19
+
20
+ ## Model description
21
+
22
+ More information needed
23
+
24
+ ## Intended uses & limitations
25
+
26
+ More information needed
27
+
28
+ ## Training and evaluation data
29
+
30
+ More information needed
31
+
32
+ ## Training procedure
33
+
34
+ ### Training hyperparameters
35
+
36
+ The following hyperparameters were used during training:
37
+ - learning_rate: 0.0001
38
+ - train_batch_size: 16
39
+ - eval_batch_size: 16
40
+ - seed: 42
41
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
+ - lr_scheduler_type: linear
43
+ - num_epochs: 10
44
+
45
+ ### Training results
46
+
47
+ | Training Loss | Epoch | Step | Validation Loss |
48
+ |:-------------:|:-----:|:-----:|:---------------:|
49
+ | 1.8478 | 1.0 | 8255 | 1.9560 |
50
+ | 1.5614 | 2.0 | 16510 | 1.8805 |
51
+ | 1.3201 | 3.0 | 24765 | 1.8681 |
52
+ | 1.1333 | 4.0 | 33020 | 2.1644 |
53
+ | 0.9384 | 5.0 | 41275 | 2.1056 |
54
+ | 0.778 | 6.0 | 49530 | 2.3509 |
55
+ | 0.6555 | 7.0 | 57785 | 2.7690 |
56
+ | 0.5564 | 8.0 | 66040 | 3.2649 |
57
+ | 0.4772 | 9.0 | 74295 | 3.7807 |
58
+ | 0.4322 | 10.0 | 82550 | 4.2753 |
59
+
60
+
61
+ ### Framework versions
62
+
63
+ - Transformers 4.38.2
64
+ - Pytorch 1.13.1+cu116
65
+ - Datasets 2.18.0
66
+ - Tokenizers 0.15.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:885dc0aa708db434257c4055981dd36e5552681be22c0636ae3bbf43475286b2
3
  size 430908208
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65b3545c72bf44c1f3119767c72388b157485910a272241ce5c0f353945ef60c
3
  size 430908208
tokenizer.json CHANGED
@@ -1,21 +1,7 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 384,
6
- "strategy": "OnlySecond",
7
- "stride": 128
8
- },
9
- "padding": {
10
- "strategy": {
11
- "Fixed": 384
12
- },
13
- "direction": "Right",
14
- "pad_to_multiple_of": null,
15
- "pad_id": 0,
16
- "pad_type_id": 0,
17
- "pad_token": "[PAD]"
18
- },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0518c512ffb1d36a4ad38e807140f22b2a48633df542555c717101f81dc122a0
3
  size 4475
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b91c69210aff88a714117c646a25b28e579cc776a13fe107df10144c0f2cc0a
3
  size 4475