alexredna commited on
Commit
521622c
1 Parent(s): a6341a9

Model save

Browse files
README.md CHANGED
@@ -1,5 +1,4 @@
1
  ---
2
- license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - trl
@@ -7,7 +6,7 @@ tags:
7
  - generated_from_trainer
8
  datasets:
9
  - generator
10
- base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
11
  model-index:
12
  - name: Tukan-1.1B-Chat-reasoning-sft
13
  results: []
@@ -18,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # Tukan-1.1B-Chat-reasoning-sft
20
 
21
- This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on the generator dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 1.2243
24
 
25
  ## Model description
26
 
@@ -54,10 +53,10 @@ The following hyperparameters were used during training:
54
 
55
  | Training Loss | Epoch | Step | Validation Loss |
56
  |:-------------:|:-----:|:----:|:---------------:|
57
- | 1.3384 | 0.24 | 10 | 1.2810 |
58
- | 1.2633 | 0.47 | 20 | 1.2418 |
59
- | 1.2495 | 0.71 | 30 | 1.2277 |
60
- | 1.2291 | 0.94 | 40 | 1.2244 |
61
 
62
 
63
  ### Framework versions
 
1
  ---
 
2
  library_name: peft
3
  tags:
4
  - trl
 
6
  - generated_from_trainer
7
  datasets:
8
  - generator
9
+ base_model: data/Tukan-1.1B-Chat-reasoning-sft_merged
10
  model-index:
11
  - name: Tukan-1.1B-Chat-reasoning-sft
12
  results: []
 
17
 
18
  # Tukan-1.1B-Chat-reasoning-sft
19
 
20
+ This model was trained from scratch on the generator dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 1.1380
23
 
24
  ## Model description
25
 
 
53
 
54
  | Training Loss | Epoch | Step | Validation Loss |
55
  |:-------------:|:-----:|:----:|:---------------:|
56
+ | 1.2298 | 0.24 | 10 | 1.1998 |
57
+ | 1.1885 | 0.47 | 20 | 1.1653 |
58
+ | 1.1674 | 0.71 | 30 | 1.1437 |
59
+ | 1.1425 | 0.94 | 40 | 1.1381 |
60
 
61
 
62
  ### Framework versions
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af46676dbe82858b87ca1547c71c35fb5c75de6b10ae4a9ef49737c41e7ae06c
3
  size 26361536
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e7920098c524c2c57950f253babaf3c8f9f48ecf07eeb9c21fd35574bf96494
3
  size 26361536
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 0.99,
3
- "eval_loss": 1.2242895364761353,
4
- "eval_runtime": 24.7608,
5
  "eval_samples": 662,
6
- "eval_samples_per_second": 5.21,
7
- "eval_steps_per_second": 1.737,
8
- "train_loss": 1.282569306237357,
9
- "train_runtime": 3226.4767,
10
  "train_samples": 25778,
11
- "train_samples_per_second": 1.576,
12
  "train_steps_per_second": 0.013
13
  }
 
1
  {
2
  "epoch": 0.99,
3
+ "eval_loss": 1.1380302906036377,
4
+ "eval_runtime": 24.4133,
5
  "eval_samples": 662,
6
+ "eval_samples_per_second": 5.284,
7
+ "eval_steps_per_second": 1.761,
8
+ "train_loss": 1.1835048993428547,
9
+ "train_runtime": 3217.2275,
10
  "train_samples": 25778,
11
+ "train_samples_per_second": 1.58,
12
  "train_steps_per_second": 0.013
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.99,
3
- "eval_loss": 1.2242895364761353,
4
- "eval_runtime": 24.7608,
5
  "eval_samples": 662,
6
- "eval_samples_per_second": 5.21,
7
- "eval_steps_per_second": 1.737
8
  }
 
1
  {
2
  "epoch": 0.99,
3
+ "eval_loss": 1.1380302906036377,
4
+ "eval_runtime": 24.4133,
5
  "eval_samples": 662,
6
+ "eval_samples_per_second": 5.284,
7
+ "eval_steps_per_second": 1.761
8
  }
runs/Jan20_09-47-06_98f107f1aa39/events.out.tfevents.1705747412.98f107f1aa39.59272.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b251b451f30eebb65f8c6f2e9add2583d9865c3ab0b2e7e8860eda8ce441ef33
3
- size 6977
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be2a18398cffadd3c9d36fe951c2e8dc0816f24d9afcb29c43d3579da8210556
3
+ size 7325
runs/Jan20_09-47-06_98f107f1aa39/events.out.tfevents.1705750653.98f107f1aa39.59272.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49057e8ddadd2ef376f5408586cca18ce2b101f1754327f775ac419cfd347940
3
+ size 354
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.99,
3
- "train_loss": 1.282569306237357,
4
- "train_runtime": 3226.4767,
5
  "train_samples": 25778,
6
- "train_samples_per_second": 1.576,
7
  "train_steps_per_second": 0.013
8
  }
 
1
  {
2
  "epoch": 0.99,
3
+ "train_loss": 1.1835048993428547,
4
+ "train_runtime": 3217.2275,
5
  "train_samples": 25778,
6
+ "train_samples_per_second": 1.58,
7
  "train_steps_per_second": 0.013
8
  }
trainer_state.json CHANGED
@@ -11,96 +11,96 @@
11
  {
12
  "epoch": 0.02,
13
  "learning_rate": 4.9930094929529506e-05,
14
- "loss": 1.458,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.12,
19
  "learning_rate": 4.827184371610511e-05,
20
- "loss": 1.4238,
21
  "step": 5
22
  },
23
  {
24
  "epoch": 0.24,
25
  "learning_rate": 4.332629679574566e-05,
26
- "loss": 1.3384,
27
  "step": 10
28
  },
29
  {
30
  "epoch": 0.24,
31
- "eval_loss": 1.2809978723526,
32
- "eval_runtime": 24.7617,
33
- "eval_samples_per_second": 5.21,
34
- "eval_steps_per_second": 1.737,
35
  "step": 10
36
  },
37
  {
38
  "epoch": 0.35,
39
  "learning_rate": 3.5847093477938956e-05,
40
- "loss": 1.2865,
41
  "step": 15
42
  },
43
  {
44
  "epoch": 0.47,
45
  "learning_rate": 2.686825233966061e-05,
46
- "loss": 1.2633,
47
  "step": 20
48
  },
49
  {
50
  "epoch": 0.47,
51
- "eval_loss": 1.2418025732040405,
52
- "eval_runtime": 24.7653,
53
- "eval_samples_per_second": 5.209,
54
- "eval_steps_per_second": 1.736,
55
  "step": 20
56
  },
57
  {
58
  "epoch": 0.59,
59
  "learning_rate": 1.7631120639727393e-05,
60
- "loss": 1.2433,
61
  "step": 25
62
  },
63
  {
64
  "epoch": 0.71,
65
  "learning_rate": 9.412754953531663e-06,
66
- "loss": 1.2495,
67
  "step": 30
68
  },
69
  {
70
  "epoch": 0.71,
71
- "eval_loss": 1.2276524305343628,
72
- "eval_runtime": 24.7715,
73
- "eval_samples_per_second": 5.208,
74
- "eval_steps_per_second": 1.736,
75
  "step": 30
76
  },
77
  {
78
  "epoch": 0.83,
79
  "learning_rate": 3.3493649053890326e-06,
80
- "loss": 1.2364,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 0.94,
85
  "learning_rate": 2.7922934437178695e-07,
86
- "loss": 1.2291,
87
  "step": 40
88
  },
89
  {
90
  "epoch": 0.94,
91
- "eval_loss": 1.2243515253067017,
92
- "eval_runtime": 24.7908,
93
- "eval_samples_per_second": 5.204,
94
- "eval_steps_per_second": 1.735,
95
  "step": 40
96
  },
97
  {
98
  "epoch": 0.99,
99
  "step": 42,
100
  "total_flos": 6.447644673468006e+16,
101
- "train_loss": 1.282569306237357,
102
- "train_runtime": 3226.4767,
103
- "train_samples_per_second": 1.576,
104
  "train_steps_per_second": 0.013
105
  }
106
  ],
 
11
  {
12
  "epoch": 0.02,
13
  "learning_rate": 4.9930094929529506e-05,
14
+ "loss": 1.2348,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.12,
19
  "learning_rate": 4.827184371610511e-05,
20
+ "loss": 1.2258,
21
  "step": 5
22
  },
23
  {
24
  "epoch": 0.24,
25
  "learning_rate": 4.332629679574566e-05,
26
+ "loss": 1.2298,
27
  "step": 10
28
  },
29
  {
30
  "epoch": 0.24,
31
+ "eval_loss": 1.1997809410095215,
32
+ "eval_runtime": 24.4233,
33
+ "eval_samples_per_second": 5.282,
34
+ "eval_steps_per_second": 1.761,
35
  "step": 10
36
  },
37
  {
38
  "epoch": 0.35,
39
  "learning_rate": 3.5847093477938956e-05,
40
+ "loss": 1.2091,
41
  "step": 15
42
  },
43
  {
44
  "epoch": 0.47,
45
  "learning_rate": 2.686825233966061e-05,
46
+ "loss": 1.1885,
47
  "step": 20
48
  },
49
  {
50
  "epoch": 0.47,
51
+ "eval_loss": 1.1652644872665405,
52
+ "eval_runtime": 24.4242,
53
+ "eval_samples_per_second": 5.282,
54
+ "eval_steps_per_second": 1.761,
55
  "step": 20
56
  },
57
  {
58
  "epoch": 0.59,
59
  "learning_rate": 1.7631120639727393e-05,
60
+ "loss": 1.1647,
61
  "step": 25
62
  },
63
  {
64
  "epoch": 0.71,
65
  "learning_rate": 9.412754953531663e-06,
66
+ "loss": 1.1674,
67
  "step": 30
68
  },
69
  {
70
  "epoch": 0.71,
71
+ "eval_loss": 1.1437482833862305,
72
+ "eval_runtime": 24.4259,
73
+ "eval_samples_per_second": 5.281,
74
+ "eval_steps_per_second": 1.76,
75
  "step": 30
76
  },
77
  {
78
  "epoch": 0.83,
79
  "learning_rate": 3.3493649053890326e-06,
80
+ "loss": 1.151,
81
  "step": 35
82
  },
83
  {
84
  "epoch": 0.94,
85
  "learning_rate": 2.7922934437178695e-07,
86
+ "loss": 1.1425,
87
  "step": 40
88
  },
89
  {
90
  "epoch": 0.94,
91
+ "eval_loss": 1.1380583047866821,
92
+ "eval_runtime": 24.4167,
93
+ "eval_samples_per_second": 5.283,
94
+ "eval_steps_per_second": 1.761,
95
  "step": 40
96
  },
97
  {
98
  "epoch": 0.99,
99
  "step": 42,
100
  "total_flos": 6.447644673468006e+16,
101
+ "train_loss": 1.1835048993428547,
102
+ "train_runtime": 3217.2275,
103
+ "train_samples_per_second": 1.58,
104
  "train_steps_per_second": 0.013
105
  }
106
  ],