000jd commited on
Commit
0c18a6b
1 Parent(s): 9bc1790

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -21,8 +21,8 @@
21
  "revision": null,
22
  "target_modules": [
23
  "k_proj",
24
- "v_proj",
25
- "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
21
  "revision": null,
22
  "target_modules": [
23
  "k_proj",
24
+ "q_proj",
25
+ "v_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f8ae74624600ac4f8e404a42663f036a9ccaa373089dcd54579c5686cd86259c
3
  size 18893616
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7689ddfff93bbceb4a0c2dff091f8df469b5e47451340e2e92098ecc18be0370
3
  size 18893616
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/phi-1_5",
3
+ "architectures": [
4
+ "PhiForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": null,
8
+ "embd_pdrop": 0.0,
9
+ "eos_token_id": null,
10
+ "hidden_act": "gelu_new",
11
+ "hidden_size": 2048,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 8192,
14
+ "layer_norm_eps": 1e-05,
15
+ "max_position_embeddings": 2048,
16
+ "model_type": "phi",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 24,
19
+ "num_key_value_heads": 32,
20
+ "partial_rotary_factor": 0.5,
21
+ "qk_layernorm": false,
22
+ "resid_pdrop": 0.0,
23
+ "rope_scaling": null,
24
+ "rope_theta": 10000.0,
25
+ "tie_word_embeddings": false,
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.41.2",
28
+ "use_cache": true,
29
+ "vocab_size": 51200
30
+ }
generation_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "transformers_version": "4.41.2"
4
+ }
runs/Jun11_02-58-40_37800b88aefd/events.out.tfevents.1718074720.37800b88aefd.16310.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77082d69aa9025781cd9690a83e530604c4e42161ff0c0810c96336d084ed394
3
+ size 7870
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1e582c0aba1eef1b2c5034c44bb0a27561d3bf4d98716f1f179575976cfc8131
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d0de1eedf61cf6fc50ad143d7e40d134c280356143c306af1dd76584132ca32a
3
  size 5112