terry69 commited on
Commit
ba537a8
1 Parent(s): 45a3b7a

Model save

Browse files
README.md CHANGED
@@ -1,16 +1,11 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: mistralai/Mistral-7B-Instruct-v0.2
4
  tags:
5
- - alignment-handbook
6
  - trl
7
  - sft
8
  - generated_from_trainer
9
- - trl
10
- - sft
11
- - generated_from_trainer
12
- datasets:
13
- - preference-data
14
  model-index:
15
  - name: feedback_p0.1_seed42_level2_syntaxmixbatch16
16
  results: []
@@ -21,9 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  # feedback_p0.1_seed42_level2_syntaxmixbatch16
23
 
24
- This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the preference-data dataset.
25
- It achieves the following results on the evaluation set:
26
- - Loss: 0.3189
27
 
28
  ## Model description
29
 
@@ -57,14 +50,11 @@ The following hyperparameters were used during training:
57
 
58
  ### Training results
59
 
60
- | Training Loss | Epoch | Step | Validation Loss |
61
- |:-------------:|:-----:|:----:|:---------------:|
62
- | 0.3656 | 1.0 | 1950 | 0.3189 |
63
 
64
 
65
  ### Framework versions
66
 
67
- - Transformers 4.43.4
68
  - Pytorch 2.3.1+cu121
69
  - Datasets 2.19.1
70
  - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: mistralai/Mistral-7B-Instruct-v0.2
5
  tags:
 
6
  - trl
7
  - sft
8
  - generated_from_trainer
 
 
 
 
 
9
  model-index:
10
  - name: feedback_p0.1_seed42_level2_syntaxmixbatch16
11
  results: []
 
16
 
17
  # feedback_p0.1_seed42_level2_syntaxmixbatch16
18
 
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on the None dataset.
 
 
20
 
21
  ## Model description
22
 
 
50
 
51
  ### Training results
52
 
 
 
 
53
 
54
 
55
  ### Framework versions
56
 
57
+ - Transformers 4.44.2
58
  - Pytorch 2.3.1+cu121
59
  - Datasets 2.19.1
60
  - Tokenizers 0.19.1
all_results.json CHANGED
@@ -1,14 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "eval_loss": 0.3188657760620117,
4
- "eval_runtime": 1.3214,
5
- "eval_samples": 10,
6
- "eval_samples_per_second": 1.514,
7
- "eval_steps_per_second": 0.757,
8
- "total_flos": 204145164288000.0,
9
- "train_loss": 0.5252774189068721,
10
- "train_runtime": 19779.9981,
11
  "train_samples": 98927,
12
- "train_samples_per_second": 1.577,
13
- "train_steps_per_second": 0.099
14
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 194513700126720.0,
4
+ "train_loss": 0.5105859521772428,
5
+ "train_runtime": 16588.6757,
 
 
 
 
 
6
  "train_samples": 98927,
7
+ "train_samples_per_second": 1.791,
8
+ "train_steps_per_second": 0.112
9
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "transformers_version": "4.43.4"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "transformers_version": "4.44.2"
6
  }
runs/Sep17_05-11-37_COE-CS-sv003/events.out.tfevents.1726550202.COE-CS-sv003.3171136.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cac4eb2784bd876a490ecb1652aa9b14fd9f654392d148486954dc85f24cb0e7
3
- size 83877
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:923f4e74a2f0b0a29dc3f113f1b7d70086baee738a395f03cb68c9c9b088092c
3
+ size 84231
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "epoch": 1.0,
3
- "total_flos": 204145164288000.0,
4
- "train_loss": 0.5252774189068721,
5
- "train_runtime": 19779.9981,
6
  "train_samples": 98927,
7
- "train_samples_per_second": 1.577,
8
- "train_steps_per_second": 0.099
9
  }
 
1
  {
2
  "epoch": 1.0,
3
+ "total_flos": 194513700126720.0,
4
+ "train_loss": 0.5105859521772428,
5
+ "train_runtime": 16588.6757,
6
  "train_samples": 98927,
7
+ "train_samples_per_second": 1.791,
8
+ "train_steps_per_second": 0.112
9
  }
trainer_state.json CHANGED
The diff for this file is too large to render. See raw diff