yuriachermann commited on
Commit
23de63f
1 Parent(s): 09a0336

Training in progress, step 500

Browse files
README.md CHANGED
@@ -5,8 +5,6 @@ tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- - ipex
9
- - GPU Max 1100
10
  base_model: google/gemma-2b
11
  datasets:
12
  - generator
@@ -36,10 +34,6 @@ More information needed
36
 
37
  More information needed
38
 
39
- ## Training Hardware
40
-
41
- This model was trained using Intel(R) Data Center GPU Max 1100
42
-
43
  ## Training procedure
44
 
45
  ### Training hyperparameters
 
5
  - trl
6
  - sft
7
  - generated_from_trainer
 
 
8
  base_model: google/gemma-2b
9
  datasets:
10
  - generator
 
34
 
35
  More information needed
36
 
 
 
 
 
37
  ## Training procedure
38
 
39
  ### Training hyperparameters
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "down_proj",
24
- "o_proj",
25
- "k_proj",
26
  "q_proj",
27
  "up_proj",
28
- "gate_proj",
29
- "v_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
24
  "down_proj",
 
 
25
  "q_proj",
26
  "up_proj",
27
+ "k_proj",
28
+ "v_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90ea7af5d7fd4c011f8a2802794d00518c27a6ec259b0546fcc48bf1e9b46eb3
3
  size 156926880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24429dd8bff3af0e92dbcb30dcf0193ff9af1d61cb4463da96e1a534f41857ac
3
  size 156926880
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48bc734926e958c820e14b3918d8edac20c3d49097f85fedbfb3b48b101d0bbb
3
  size 4539
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b634a18d51cacfbd1b62bdb9139a2391ed2c881d11bf868f93ff2e14b12e33ee
3
  size 4539