AmberYifan commited on
Commit
d3bda95
1 Parent(s): 5c51859

End of training

Browse files
README.md CHANGED
@@ -1,9 +1,15 @@
1
  ---
2
  base_model: argsearch/llama-7b-sft-float32
3
  tags:
 
4
  - trl
5
  - dpo
6
  - generated_from_trainer
 
 
 
 
 
7
  model-index:
8
  - name: llama-7b-sft-DPO
9
  results: []
@@ -14,17 +20,17 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # llama-7b-sft-DPO
16
 
17
- This model is a fine-tuned version of [argsearch/llama-7b-sft-float32](https://huggingface.co/argsearch/llama-7b-sft-float32) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
  - Loss: 0.6525
20
- - Rewards/chosen: 0.3312
21
- - Rewards/rejected: 0.1950
22
- - Rewards/accuracies: 0.6074
23
- - Rewards/margins: 0.1363
24
- - Logps/rejected: -633.4115
25
- - Logps/chosen: -690.5892
26
- - Logits/rejected: -1.9098
27
- - Logits/chosen: -1.9645
28
 
29
  ## Model description
30
 
 
1
  ---
2
  base_model: argsearch/llama-7b-sft-float32
3
  tags:
4
+ - alignment-handbook
5
  - trl
6
  - dpo
7
  - generated_from_trainer
8
+ - trl
9
+ - dpo
10
+ - generated_from_trainer
11
+ datasets:
12
+ - Dahoas/full-hh-rlhf
13
  model-index:
14
  - name: llama-7b-sft-DPO
15
  results: []
 
20
 
21
  # llama-7b-sft-DPO
22
 
23
+ This model is a fine-tuned version of [argsearch/llama-7b-sft-float32](https://huggingface.co/argsearch/llama-7b-sft-float32) on the Dahoas/full-hh-rlhf dataset.
24
  It achieves the following results on the evaluation set:
25
  - Loss: 0.6525
26
+ - Rewards/chosen: 0.3315
27
+ - Rewards/rejected: 0.1953
28
+ - Rewards/accuracies: 0.6080
29
+ - Rewards/margins: 0.1362
30
+ - Logps/rejected: -633.3815
31
+ - Logps/chosen: -690.5654
32
+ - Logits/rejected: -1.9212
33
+ - Logits/chosen: -1.9766
34
 
35
  ## Model description
36
 
all_results.json CHANGED
@@ -1,5 +1,18 @@
1
  {
2
  "epoch": 1.0,
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  "train_loss": 0.6642669952984608,
4
  "train_runtime": 38056.9659,
5
  "train_samples": 112052,
 
1
  {
2
  "epoch": 1.0,
3
+ "eval_logits/chosen": -1.97661554813385,
4
+ "eval_logits/rejected": -1.9212419986724854,
5
+ "eval_logps/chosen": -690.5653686523438,
6
+ "eval_logps/rejected": -633.3815307617188,
7
+ "eval_loss": 0.6524845957756042,
8
+ "eval_rewards/accuracies": 0.6080127954483032,
9
+ "eval_rewards/chosen": 0.33147066831588745,
10
+ "eval_rewards/margins": 0.13620509207248688,
11
+ "eval_rewards/rejected": 0.19526559114456177,
12
+ "eval_runtime": 812.3144,
13
+ "eval_samples": 12451,
14
+ "eval_samples_per_second": 15.328,
15
+ "eval_steps_per_second": 0.48,
16
  "train_loss": 0.6642669952984608,
17
  "train_runtime": 38056.9659,
18
  "train_samples": 112052,
config.json CHANGED
@@ -25,6 +25,6 @@
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.39.0.dev0",
28
- "use_cache": false,
29
  "vocab_size": 32000
30
  }
 
25
  "tie_word_embeddings": false,
26
  "torch_dtype": "bfloat16",
27
  "transformers_version": "4.39.0.dev0",
28
+ "use_cache": true,
29
  "vocab_size": 32000
30
  }
eval_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 1.0,
3
+ "eval_logits/chosen": -1.97661554813385,
4
+ "eval_logits/rejected": -1.9212419986724854,
5
+ "eval_logps/chosen": -690.5653686523438,
6
+ "eval_logps/rejected": -633.3815307617188,
7
+ "eval_loss": 0.6524845957756042,
8
+ "eval_rewards/accuracies": 0.6080127954483032,
9
+ "eval_rewards/chosen": 0.33147066831588745,
10
+ "eval_rewards/margins": 0.13620509207248688,
11
+ "eval_rewards/rejected": 0.19526559114456177,
12
+ "eval_runtime": 812.3144,
13
+ "eval_samples": 12451,
14
+ "eval_samples_per_second": 15.328,
15
+ "eval_steps_per_second": 0.48
16
+ }
runs/Apr30_13-22-22_gilbreth-j001.rcac.purdue.edu/events.out.tfevents.1714537921.gilbreth-j001.rcac.purdue.edu.163553.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e82a4de46062d6569253e1aa1860ad7cf161bb1933cc15ea2a36262d673b4ee
3
+ size 828