keylazy commited on
Commit
f6a0f7b
1 Parent(s): 512e645

Training in progress, epoch 0

Browse files
config.json CHANGED
@@ -6,15 +6,15 @@
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
8
  "hidden_act": "silu",
9
- "hidden_size": 128,
10
  "initializer_range": 0.02,
11
- "intermediate_size": 256,
12
  "max_position_embeddings": 2048,
13
  "model_type": "llama",
14
- "num_attention_heads": 8,
15
- "num_hidden_layers": 3,
16
- "num_key_value_heads": 8,
17
- "pad_token_id": 2,
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
@@ -23,5 +23,5 @@
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.35.0",
25
  "use_cache": true,
26
- "vocab_size": 5000
27
  }
 
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
8
  "hidden_act": "silu",
9
+ "hidden_size": 768,
10
  "initializer_range": 0.02,
11
+ "intermediate_size": 2304,
12
  "max_position_embeddings": 2048,
13
  "model_type": "llama",
14
+ "num_attention_heads": 12,
15
+ "num_hidden_layers": 12,
16
+ "num_key_value_heads": 12,
17
+ "pad_token_id": 3,
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-06,
20
  "rope_scaling": null,
 
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.35.0",
25
  "use_cache": true,
26
+ "vocab_size": 10000
27
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a44e7d6e5b528530ea499c1816d26d896b51582f02e86fdf1ef2603247e5dece
3
- size 7092848
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:933b85c55b2f90ec70e21aa648b81a66d44a9210da1748206a7fd66fd14cbe00
3
+ size 429579272
runs/Nov10_08-41-42_ddf776d2631b/events.out.tfevents.1699605803.ddf776d2631b.2960.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60ee64d30c5ff92e1db4f2fb7e742c1e5baaf787318d1fa3b78048c3813f1a1b
3
+ size 4359
runs/Nov10_08-52-09_ddf776d2631b/events.out.tfevents.1699606339.ddf776d2631b.5801.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef14827bc46ef6b77a28c0b6f3fef85c7c38c88f6d4f51dfdcf1545005c1d3a
3
+ size 4787
special_tokens_map.json CHANGED
@@ -13,7 +13,13 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<pad>",
 
 
 
 
 
 
17
  "unk_token": {
18
  "content": "<unk>",
19
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<pad>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f283a54aa10902c1be4fea3071a0df9998e466bdab8c4d215c10c34de662920b
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c44fcd57369efc5b4f98de198677d74728c8f86c0e43ba5a8469346ace4f1b3f
3
  size 4600