taicheng commited on
Commit
f77f5fa
1 Parent(s): f8905d8

Model save

Browse files
README.md CHANGED
@@ -3,15 +3,10 @@ library_name: transformers
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
6
- - alignment-handbook
7
- - trl
8
- - dpo
9
- - generated_from_trainer
10
  - trl
11
  - dpo
 
12
  - generated_from_trainer
13
- datasets:
14
- - HuggingFaceH4/ultrafeedback_binarized
15
  model-index:
16
  - name: zephyr-7b-align-scan-0.0-0.0-cosine-2
17
  results: []
@@ -22,17 +17,17 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  # zephyr-7b-align-scan-0.0-0.0-cosine-2
24
 
25
- This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on the HuggingFaceH4/ultrafeedback_binarized dataset.
26
  It achieves the following results on the evaluation set:
27
- - Loss: 0.6691
28
- - Rewards/chosen: -0.3145
29
- - Rewards/rejected: -0.5033
30
- - Rewards/accuracies: 0.3075
31
- - Rewards/margins: 0.1888
32
- - Logps/rejected: -601.5961
33
- - Logps/chosen: -399.7386
34
- - Logits/rejected: 4.1594
35
- - Logits/chosen: 3.9982
36
 
37
  ## Model description
38
 
@@ -51,7 +46,7 @@ More information needed
51
  ### Training hyperparameters
52
 
53
  The following hyperparameters were used during training:
54
- - learning_rate: 8.932373406244385e-07
55
  - train_batch_size: 8
56
  - eval_batch_size: 8
57
  - seed: 42
@@ -67,9 +62,9 @@ The following hyperparameters were used during training:
67
 
68
  ### Training results
69
 
70
- | Training Loss | Epoch | Step | Validation Loss | Rewards/chosen | Rewards/rejected | Rewards/accuracies | Rewards/margins | Logps/rejected | Logps/chosen | Logits/rejected | Logits/chosen |
71
- |:-------------:|:------:|:----:|:---------------:|:--------------:|:----------------:|:------------------:|:---------------:|:--------------:|:------------:|:---------------:|:-------------:|
72
- | 0.66 | 1.0417 | 100 | 0.6710 | -0.1518 | -0.2380 | 0.3155 | 0.0862 | -327.2254 | -231.4086 | 2.3998 | 2.2227 |
73
 
74
 
75
  ### Framework versions
 
3
  license: apache-2.0
4
  base_model: alignment-handbook/zephyr-7b-sft-full
5
  tags:
 
 
 
 
6
  - trl
7
  - dpo
8
+ - alignment-handbook
9
  - generated_from_trainer
 
 
10
  model-index:
11
  - name: zephyr-7b-align-scan-0.0-0.0-cosine-2
12
  results: []
 
17
 
18
  # zephyr-7b-align-scan-0.0-0.0-cosine-2
19
 
20
+ This model is a fine-tuned version of [alignment-handbook/zephyr-7b-sft-full](https://huggingface.co/alignment-handbook/zephyr-7b-sft-full) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Logits/chosen: 2.2227
23
+ - Logits/rejected: 2.3998
24
+ - Logps/chosen: -231.4086
25
+ - Logps/rejected: -327.2254
26
+ - Loss: 0.6710
27
+ - Rewards/accuracies: 0.3155
28
+ - Rewards/chosen: -0.1518
29
+ - Rewards/margins: 0.0862
30
+ - Rewards/rejected: -0.2380
31
 
32
  ## Model description
33
 
 
46
  ### Training hyperparameters
47
 
48
  The following hyperparameters were used during training:
49
+ - learning_rate: 3.454537113325051e-07
50
  - train_batch_size: 8
51
  - eval_batch_size: 8
52
  - seed: 42
 
62
 
63
  ### Training results
64
 
65
+ | Training Loss | Epoch | Step | Logits/chosen | Logits/rejected | Logps/chosen | Logps/rejected | Validation Loss | Rewards/accuracies | Rewards/chosen | Rewards/margins | Rewards/rejected |
66
+ |:-------------:|:------:|:----:|:-------------:|:---------------:|:------------:|:--------------:|:---------------:|:------------------:|:--------------:|:---------------:|:----------------:|
67
+ | 0.66 | 1.0417 | 100 | 2.2227 | 2.3998 | -231.4086 | -327.2254 | 0.6710 | 0.3155 | -0.1518 | 0.0862 | -0.2380 |
68
 
69
 
70
  ### Framework versions
all_results.json CHANGED
@@ -14,9 +14,9 @@
14
  "eval_samples_per_second": 17.613,
15
  "eval_steps_per_second": 0.555,
16
  "total_flos": 0.0,
17
- "train_loss": 0.6542075766871372,
18
- "train_runtime": 2362.1376,
19
  "train_samples": 6113,
20
- "train_samples_per_second": 5.176,
21
- "train_steps_per_second": 0.081
22
  }
 
14
  "eval_samples_per_second": 17.613,
15
  "eval_steps_per_second": 0.555,
16
  "total_flos": 0.0,
17
+ "train_loss": 0.0,
18
+ "train_runtime": 0.0151,
19
  "train_samples": 6113,
20
+ "train_samples_per_second": 807844.742,
21
+ "train_steps_per_second": 12686.585
22
  }
config.json CHANGED
@@ -22,6 +22,6 @@
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
- "use_cache": true,
26
  "vocab_size": 32000
27
  }
 
22
  "tie_word_embeddings": false,
23
  "torch_dtype": "bfloat16",
24
  "transformers_version": "4.44.2",
25
+ "use_cache": false,
26
  "vocab_size": 32000
27
  }
runs/Sep28_05-41-43_qa-a100-005.crc.nd.edu/events.out.tfevents.1727516750.qa-a100-005.crc.nd.edu.103038.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20e9a72ad1adfba4d325330f4f84d263ce178b64b3dd7c8d3ff59bce3cd57385
3
+ size 6808
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 2.0,
3
  "total_flos": 0.0,
4
- "train_loss": 0.6542075766871372,
5
- "train_runtime": 2362.1376,
6
  "train_samples": 6113,
7
- "train_samples_per_second": 5.176,
8
- "train_steps_per_second": 0.081
9
  }
 
1
  {
2
  "epoch": 2.0,
3
  "total_flos": 0.0,
4
+ "train_loss": 0.0,
5
+ "train_runtime": 0.0151,
6
  "train_samples": 6113,
7
+ "train_samples_per_second": 807844.742,
8
+ "train_steps_per_second": 12686.585
9
  }
trainer_state.json CHANGED
@@ -328,10 +328,10 @@
328
  "epoch": 2.0,
329
  "step": 192,
330
  "total_flos": 0.0,
331
- "train_loss": 0.6542075766871372,
332
- "train_runtime": 2362.1376,
333
- "train_samples_per_second": 5.176,
334
- "train_steps_per_second": 0.081
335
  }
336
  ],
337
  "logging_steps": 10,
 
328
  "epoch": 2.0,
329
  "step": 192,
330
  "total_flos": 0.0,
331
+ "train_loss": 0.0,
332
+ "train_runtime": 0.0151,
333
+ "train_samples_per_second": 807844.742,
334
+ "train_steps_per_second": 12686.585
335
  }
336
  ],
337
  "logging_steps": 10,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:62f8749eb732d929518f27ac38e596a1f235270396c09fc8453f000ed47f4d13
3
  size 7672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e72b12738e7223504a07d5d00dad20ac1780fd7771a882714ca5d69e9ccd7ee
3
  size 7672