Intradiction commited on
Commit
68f5e9e
1 Parent(s): 2915bdc

Training in progress, epoch 1

Browse files
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "bert-base-uncased",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -10,7 +10,7 @@
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
  "lora_alpha": 1,
13
- "lora_dropout": 0.1,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
@@ -19,8 +19,8 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "query",
23
- "value"
24
  ],
25
  "task_type": "SEQ_CLS",
26
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
  "lora_alpha": 1,
13
+ "lora_dropout": 0.01,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
+ "value",
23
+ "query"
24
  ],
25
  "task_type": "SEQ_CLS",
26
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e44ce263e6fd885f50d82ca515b9325375b43ee36ededb75acf161ce88bc2e41
3
- size 48
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dae7262ece0c91e504aac397e110aedd5e50c37830e186c6d7c9392166c30e5
3
+ size 161256
runs/Mar15_15-42-57_c9face729cf4/events.out.tfevents.1710517378.c9face729cf4.247.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8ee626b673d0c6b8f67843e21194b07d5f71285b03fb70a036a580b4552cfefe
3
+ size 4665
runs/Mar15_15-44-00_c9face729cf4/events.out.tfevents.1710517441.c9face729cf4.247.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36765cdbeb7ae0ff0c2838c0a8e724944e1893519a7c2545fa5892e110175a52
3
+ size 4665
runs/Mar15_15-45-44_c9face729cf4/events.out.tfevents.1710517546.c9face729cf4.247.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3cddf30e7af983b0cac53e69f5a2af97ad6dc1f1dcd0b4c6459b77c6c102321
3
+ size 4980
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34cea84ac4052828d70a822a1fdc04f50474c27092f277d2681af733008e1790
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bd45b9cd1b7cfe4e781be6834c516dbf323bb3ecd21271bc3599d21983ed695
3
  size 4920