lvcalucioli commited on
Commit
7c024ca
1 Parent(s): faef472

llamantino7b_2_question-answering

Browse files
README.md CHANGED
@@ -42,7 +42,7 @@ The following hyperparameters were used during training:
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
- - num_epochs: 5
46
 
47
  ### Training results
48
 
@@ -51,7 +51,7 @@ The following hyperparameters were used during training:
51
  ### Framework versions
52
 
53
  - PEFT 0.8.2
54
- - Transformers 4.38.0.dev0
55
  - Pytorch 2.0.1+cu117
56
  - Datasets 2.16.1
57
  - Tokenizers 0.15.2
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
45
+ - num_epochs: 14
46
 
47
  ### Training results
48
 
 
51
  ### Framework versions
52
 
53
  - PEFT 0.8.2
54
+ - Transformers 4.39.0.dev0
55
  - Pytorch 2.0.1+cu117
56
  - Datasets 2.16.1
57
  - Tokenizers 0.15.2
adapter_config.json CHANGED
@@ -19,14 +19,15 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
- "k_proj",
24
- "down_proj",
25
  "q_proj",
26
- "v_proj",
 
27
  "up_proj",
28
- "gate_proj"
 
 
29
  ],
30
  "task_type": "CAUSAL_LM",
 
31
  "use_rslora": false
32
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "q_proj",
23
+ "gate_proj",
24
+ "down_proj",
25
  "up_proj",
26
+ "k_proj",
27
+ "o_proj",
28
+ "v_proj"
29
  ],
30
  "task_type": "CAUSAL_LM",
31
+ "use_dora": false,
32
  "use_rslora": false
33
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2607c5f230285c3c1c9175320dbc3ea51fb6ab1a5f1f4419832fbc72c854fbc1
3
  size 639691872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc5f825d4cdee80e1f3ac5c8aab194d2aaaa5a7d66d1ddab9b3c31616fd6ad06
3
  size 639691872
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 250,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 700,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a6579f09f6bad01c6bab89da6134e31b132cc362e1e7afafcdf94b3fe9eded66
3
- size 4347
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:039ee6bcc800f9ccf323738be238b42dc6b00907ba7be0d99f0347e75dfa4c9c
3
+ size 4539