cmn commited on
Commit
7135a29
1 Parent(s): ed7a729

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -20,8 +20,8 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "v_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
  ],
26
  "task_type": "CAUSAL_LM",
27
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3233c7a4fb9514970abdbebc1beabe021c6456af4c37021697028efdc901f320
3
  size 2319484776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43fd4c67adbbfa9224dff3be65f5cad93577e3612ae45b00ca411d1e0ae796f0
3
  size 2319484776
config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 128000,
9
+ "eos_token_id": 128009,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 4096,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 14336,
14
+ "max_position_embeddings": 8192,
15
+ "mlp_bias": false,
16
+ "model_type": "llama",
17
+ "num_attention_heads": 32,
18
+ "num_hidden_layers": 32,
19
+ "num_key_value_heads": 8,
20
+ "pretraining_tp": 1,
21
+ "quantization_config": {
22
+ "_load_in_4bit": true,
23
+ "_load_in_8bit": false,
24
+ "bnb_4bit_compute_dtype": "bfloat16",
25
+ "bnb_4bit_quant_storage": "uint8",
26
+ "bnb_4bit_quant_type": "nf4",
27
+ "bnb_4bit_use_double_quant": true,
28
+ "llm_int8_enable_fp32_cpu_offload": false,
29
+ "llm_int8_has_fp16_weight": false,
30
+ "llm_int8_skip_modules": null,
31
+ "llm_int8_threshold": 6.0,
32
+ "load_in_4bit": true,
33
+ "load_in_8bit": false,
34
+ "quant_method": "bitsandbytes"
35
+ },
36
+ "rms_norm_eps": 1e-05,
37
+ "rope_scaling": null,
38
+ "rope_theta": 500000.0,
39
+ "tie_word_embeddings": false,
40
+ "torch_dtype": "bfloat16",
41
+ "transformers_version": "4.41.1",
42
+ "use_cache": true,
43
+ "vocab_size": 128257
44
+ }
generation_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 128000,
3
+ "do_sample": true,
4
+ "eos_token_id": [
5
+ 128001,
6
+ 128009
7
+ ],
8
+ "max_length": 4096,
9
+ "temperature": 0.6,
10
+ "top_p": 0.9,
11
+ "transformers_version": "4.41.1"
12
+ }
model-00001-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d350cc356a2b0dc278fc9de63535809ab8ff2ab249fb13de7a06dffb51abc7e
3
+ size 4870738467
model-00002-of-00002.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fe6254c7d2a6440c7c6ea3a6b9e89673e8fd28cdb42734d31ff1978f274e2cf
3
+ size 1050681472
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
runs/Jun03_11-13-06_lymphoma/events.out.tfevents.1717413206.lymphoma CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f916c63f61b1281544b32e1d62401855afd6905845057eeeb257a0807b138901
3
- size 5292
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3761c40ac5c07de705fce15e2af408c49949b088efca962aa43298e6c868bf63
3
+ size 5847
runs/Jun03_11-20-27_lymphoma/events.out.tfevents.1717413647.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a9fa598d9d0a2c34dc1e82f906bed57cf18867a886843e20e49279efb99970c
3
+ size 5847
runs/Jun03_11-24-18_lymphoma/events.out.tfevents.1717413877.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24e33ad1186e18da4e810b42dbfce75373d68e7489ff4dcf617cc182a4857a60
3
+ size 5292
runs/Jun03_11-27-26_lymphoma/events.out.tfevents.1717414066.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:640251821f930b7462f1632e3033f938256cf0be24c43c6cdeee4ad868539840
3
+ size 5847
runs/Jun03_11-31-35_lymphoma/events.out.tfevents.1717414314.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0a652a6f43ee9a7f862a24cd49d40a341b5760db8885e768a903882723a0e53
3
+ size 4013
runs/Jun03_11-37-10_lymphoma/events.out.tfevents.1717414650.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0de312730959c6f73b154436176021b88613aaf32d7cdbd36d9eb2d839822c61
3
+ size 4013
runs/Jun03_11-40-16_lymphoma/events.out.tfevents.1717414836.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:050d58481e66ea06c00a022083a88068821b59149edd27afe96453a2a4fa2d58
3
+ size 4013
runs/Jun03_11-45-24_lymphoma/events.out.tfevents.1717415144.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b59d43b49229a3198ecd1885baeff3b2785098e91d5e33aa714c14b19e0919ad
3
+ size 4013
runs/Jun03_11-48-55_lymphoma/events.out.tfevents.1717415354.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33f8cd76b9deb49c83e46a72a621f8b177a021678d7a9ee8d4c9b4c3c513079a
3
+ size 4013
runs/Jun03_11-51-00_lymphoma/events.out.tfevents.1717415480.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e5f18388b56f20b2ad603ad9e7c8f72301923251c550a00cef97f1d68db7b7d
3
+ size 4013
runs/Jun03_11-55-24_lymphoma/events.out.tfevents.1717415745.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:360a8d7a9ccb79d03a80daa81b3b1d90bb01c69231ef737397633633dcc86599
3
+ size 4013
runs/Jun03_11-57-53_lymphoma/events.out.tfevents.1717415893.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf9772a6dbf4e0a9c89f95f49a1e0f82916a3f68ea3a1efd211d85055b0022f7
3
+ size 4013
runs/Jun03_12-00-00_lymphoma/events.out.tfevents.1717416020.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d1123bb1c02951a3b866c9c4af185f809c557d5f2570bd5d032c5fcb567eb93
3
+ size 4013
runs/Jun03_12-09-08_lymphoma/events.out.tfevents.1717416568.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92b14dac6216494e8217a19a94fd93821efb89a5fb641cdb057301bfdb8f1f8f
3
+ size 4013
runs/Jun03_14-00-01_lymphoma/events.out.tfevents.1717423221.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a9801f48dc402778af1ed9d52a90669bdb1be53b3fe6a62dd72e91f614e93e5
3
+ size 8190
runs/Jun03_14-08-29_lymphoma/events.out.tfevents.1717423729.lymphoma ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2ba6214c88f5c52c691970ba18aeb8ca2e7fdc8cfbd37b92c9561a33d37dc97
3
+ size 23059
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e6fa487bb1e3e585834254531c7c5f8ae584bcb39612eef868e1a6f696cac54
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b485833c6a3456a1246c6054f4d47beb4e627516f4772f1153b2acd3bbdb5a0
3
  size 5112