ananyarn commited on
Commit
2e138d8
1 Parent(s): 81c0a0b

Training in progress, step 50

Browse files
adapter_config.json CHANGED
@@ -19,16 +19,15 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "down_proj",
23
  "v_proj",
 
24
  "k_proj",
 
25
  "o_proj",
26
- "up_proj",
27
  "gate_proj",
28
- "lm_head",
29
- "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
- "use_dora": false,
33
  "use_rslora": false
34
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "v_proj",
23
+ "down_proj",
24
  "k_proj",
25
+ "lm_head",
26
  "o_proj",
 
27
  "gate_proj",
28
+ "q_proj",
29
+ "up_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
 
32
  "use_rslora": false
33
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fec2d457562d49bc484cec68c6af7f353b39645eec6bdfdbb666cf2298f9ee26
3
  size 600059184
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cc53ae6dc65ec9d2397fa336c346a12d90367eef2459bf163299422571ed2c6
3
  size 600059184
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 512,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1063838239da656aacfa9f24b133c884eb1529130042f6a57077d51682e791d
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4081863c3b78c3896dad8368051fd6136fb7e1526cc288e6e99a637ef390343a
3
+ size 4664