sohui commited on
Commit
17d831d
1 Parent(s): c112f38

Training in progress, step 200

Browse files
adapter_config.json CHANGED
@@ -16,9 +16,9 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "lm_head",
20
  "c_attn",
21
- "c_proj"
22
  ],
23
  "task_type": "CAUSAL_LM"
24
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "c_proj",
20
  "c_attn",
21
+ "lm_head"
22
  ],
23
  "task_type": "CAUSAL_LM"
24
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56b4b08ffa23926ff702bcbb17c54b51e724b98235967433a0a9dacd48e42b7b
3
  size 3259600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7cc6d16cbd22cded96d7e3b41c6a4cb0bb3b5e843c7035e47a5ae7d125f91b08
3
  size 3259600
logs/events.out.tfevents.1700837147.67976616c584.819.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:789eccc24481e359c2fbd7c6acda9c07fc8a11496866dc7e40872f4f2319fec5
3
+ size 5337
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 1024,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 500,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b957b89c50af7bc913086d4ddb5f61efe5e0c4cbe76faca27783282c3ddf3849
3
- size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c25bf232e272a3a37c9069671d88753010ee865b54511bebc7604a56ac0d0713
3
+ size 4536