hynky HF staff commited on
Commit
add5963
1 Parent(s): e712931

Model save

Browse files
README.md CHANGED
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [codellama/Codellama-7b-Instruct-hf](https://huggingface.co/codellama/Codellama-7b-Instruct-hf) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 1.2840
19
 
20
  ## Model description
21
 
@@ -34,19 +34,20 @@ More information needed
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
- - learning_rate: 2e-05
38
  - train_batch_size: 4
39
  - eval_batch_size: 4
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
 
43
  - num_epochs: 1
44
 
45
  ### Training results
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
- | 1.2727 | 0.33 | 1 | 1.2840 |
50
 
51
 
52
  ### Framework versions
 
15
 
16
  This model is a fine-tuned version of [codellama/Codellama-7b-Instruct-hf](https://huggingface.co/codellama/Codellama-7b-Instruct-hf) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 1.3468
19
 
20
  ## Model description
21
 
 
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
+ - learning_rate: 4e-05
38
  - train_batch_size: 4
39
  - eval_batch_size: 4
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: cosine
43
+ - lr_scheduler_warmup_ratio: 0.1
44
  - num_epochs: 1
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss |
49
  |:-------------:|:-----:|:----:|:---------------:|
50
+ | 1.2727 | 0.33 | 1 | 1.3468 |
51
 
52
 
53
  ### Framework versions
adapter_config.json CHANGED
@@ -16,8 +16,8 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
  "q_proj",
 
21
  "k_proj",
22
  "o_proj"
23
  ],
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "q_proj",
20
+ "v_proj",
21
  "k_proj",
22
  "o_proj"
23
  ],
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41d33eb721c35016c482bda01bcbf43ff89ca011daf13361fbecfd531cceb071
3
  size 536906096
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:64d0a7e68c76899707f02f5a7ccd4909792e2a1072e068f5224c512393092796
3
  size 536906096
all_results.json CHANGED
@@ -1,13 +1,13 @@
1
  {
2
  "epoch": 0.33,
3
- "eval_loss": 1.2840009927749634,
4
- "eval_runtime": 0.5015,
5
  "eval_samples": 10,
6
- "eval_samples_per_second": 19.941,
7
- "eval_steps_per_second": 5.982,
8
  "train_loss": 1.2727434635162354,
9
- "train_runtime": 4.5773,
10
  "train_samples": 10,
11
- "train_samples_per_second": 2.185,
12
- "train_steps_per_second": 0.655
13
  }
 
1
  {
2
  "epoch": 0.33,
3
+ "eval_loss": 1.346835732460022,
4
+ "eval_runtime": 0.7097,
5
  "eval_samples": 10,
6
+ "eval_samples_per_second": 14.09,
7
+ "eval_steps_per_second": 4.227,
8
  "train_loss": 1.2727434635162354,
9
+ "train_runtime": 670.5207,
10
  "train_samples": 10,
11
+ "train_samples_per_second": 0.015,
12
+ "train_steps_per_second": 0.004
13
  }
eval_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.33,
3
- "eval_loss": 1.2840009927749634,
4
- "eval_runtime": 0.5015,
5
  "eval_samples": 10,
6
- "eval_samples_per_second": 19.941,
7
- "eval_steps_per_second": 5.982
8
  }
 
1
  {
2
  "epoch": 0.33,
3
+ "eval_loss": 1.346835732460022,
4
+ "eval_runtime": 0.7097,
5
  "eval_samples": 10,
6
+ "eval_samples_per_second": 14.09,
7
+ "eval_steps_per_second": 4.227
8
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "epoch": 0.33,
3
  "train_loss": 1.2727434635162354,
4
- "train_runtime": 4.5773,
5
  "train_samples": 10,
6
- "train_samples_per_second": 2.185,
7
- "train_steps_per_second": 0.655
8
  }
 
1
  {
2
  "epoch": 0.33,
3
  "train_loss": 1.2727434635162354,
4
+ "train_runtime": 670.5207,
5
  "train_samples": 10,
6
+ "train_samples_per_second": 0.015,
7
+ "train_steps_per_second": 0.004
8
  }
trainer_state.json CHANGED
@@ -10,16 +10,16 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.33,
13
- "learning_rate": 1.5000000000000002e-05,
14
  "loss": 1.2727,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.33,
19
- "eval_loss": 1.2840009927749634,
20
- "eval_runtime": 0.5018,
21
- "eval_samples_per_second": 19.93,
22
- "eval_steps_per_second": 5.979,
23
  "step": 1
24
  },
25
  {
@@ -27,9 +27,9 @@
27
  "step": 1,
28
  "total_flos": 82841111101440.0,
29
  "train_loss": 1.2727434635162354,
30
- "train_runtime": 4.5773,
31
- "train_samples_per_second": 2.185,
32
- "train_steps_per_second": 0.655
33
  }
34
  ],
35
  "logging_steps": 5,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.33,
13
+ "learning_rate": 4e-05,
14
  "loss": 1.2727,
15
  "step": 1
16
  },
17
  {
18
  "epoch": 0.33,
19
+ "eval_loss": 1.346835732460022,
20
+ "eval_runtime": 664.8795,
21
+ "eval_samples_per_second": 0.015,
22
+ "eval_steps_per_second": 0.005,
23
  "step": 1
24
  },
25
  {
 
27
  "step": 1,
28
  "total_flos": 82841111101440.0,
29
  "train_loss": 1.2727434635162354,
30
+ "train_runtime": 670.5207,
31
+ "train_samples_per_second": 0.015,
32
+ "train_steps_per_second": 0.004
33
  }
34
  ],
35
  "logging_steps": 5,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dab2feac7f8ecb4382b8a17fe51c3a88a2fde01e4e90be87479eb2a8d1a4a13
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87028a0aba3a915e288cd845e6c15267c332eb143e08585afac1725b7cf81cc7
3
  size 4728