Michelvh commited on
Commit
d3f5d93
1 Parent(s): 0e3be0c

qlora-llama2-7b-question-generation-eduqg

Browse files
Files changed (5) hide show
  1. README.md +5 -5
  2. adapter_config.json +2 -2
  3. adapter_model.bin +1 -1
  4. tokenizer.json +6 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -32,14 +32,14 @@ More information needed
32
 
33
  The following hyperparameters were used during training:
34
  - learning_rate: 5e-05
35
- - train_batch_size: 1
36
- - eval_batch_size: 1
37
  - seed: 42
38
- - gradient_accumulation_steps: 4
39
- - total_train_batch_size: 4
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
- - num_epochs: 9
43
 
44
  ### Training results
45
 
 
32
 
33
  The following hyperparameters were used during training:
34
  - learning_rate: 5e-05
35
+ - train_batch_size: 4
36
+ - eval_batch_size: 4
37
  - seed: 42
38
+ - gradient_accumulation_steps: 16
39
+ - total_train_batch_size: 64
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: linear
42
+ - num_epochs: 10
43
 
44
  ### Training results
45
 
adapter_config.json CHANGED
@@ -16,8 +16,8 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
- "q_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
20
+ "v_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9026fa5734c7d8e195d928191eb11d32d5e05eb8455bc9bdc176dca77273c000
3
  size 16822989
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:913e8254c2e0e7fab8dd0cb3902e6ccc76bdcc1a7eeb407c84e312218bb603a0
3
  size 16822989
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 1024,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:44ce4fd3212f03b378c2fd5c40f6ddeffb9fba512eb8f1a24e427ac892db9ec5
3
  size 4155
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20267de5e18ec40118e0e493cce3e4f294bde9ee71fb5c53154ae0351e9afdb9
3
  size 4155