shanjiaz commited on
Commit
28c5a6a
·
verified ·
1 Parent(s): 36f9404

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -1,10 +1,25 @@
1
  {
2
- "model_type": "test_tiny_model",
3
- "vocab_size": 1000,
 
 
 
 
4
  "hidden_size": 128,
 
5
  "intermediate_size": 256,
6
- "num_hidden_layers": 1,
7
- "architectures": [
8
- "TinyTestModel"
9
- ]
 
 
 
 
 
 
 
 
 
 
10
  }
 
1
  {
2
+ "architectures": [
3
+ "LlamaForCausalLM"
4
+ ],
5
+ "bos_token_id": 1,
6
+ "eos_token_id": 2,
7
+ "hidden_act": "silu",
8
  "hidden_size": 128,
9
+ "initializer_range": 0.02,
10
  "intermediate_size": 256,
11
+ "max_position_embeddings": 512,
12
+ "model_type": "llama",
13
+ "num_attention_heads": 4,
14
+ "num_hidden_layers": 2,
15
+ "num_key_value_heads": 4,
16
+ "pretraining_tp": 1,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_scaling": null,
19
+ "rope_theta": 10000.0,
20
+ "tie_word_embeddings": false,
21
+ "torch_dtype": "float32",
22
+ "transformers_version": "4.36.0",
23
+ "use_cache": true,
24
+ "vocab_size": 1000
25
  }
generation_config.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 1,
3
+ "eos_token_id": 2,
4
+ "max_length": 512
5
+ }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8523a121065d512d58f624e0f366e4bf88e1fa0655a34a4562bd8b32c518c32
3
- size 775064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a3f6296e3d9c8e0b2cb50461d2701147946359d09f476f873a4209192931f641
3
+ size 1169432
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7b6da2e1a2a6ef5ec681f1a75263ca86f2c0d4fe8ebc42a52d2ffe7f17e84136
3
- size 643856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5487fb43332d62b17d6389ed65ce9eca9ffe1def0135a9fdb1a5c025d6ddb8ac
3
+ size 1170016
model.safetensors.index.json CHANGED
@@ -1,14 +1,28 @@
1
  {
2
  "metadata": {
3
- "total_size": 1000000
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
7
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
 
 
 
 
8
  "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
9
  "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
10
- "model.layers.0.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
11
- "lm_head.weight": "model-00002-of-00002.safetensors",
12
- "model.norm.weight": "model-00002-of-00002.safetensors"
 
 
 
 
 
 
 
 
 
 
13
  }
14
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 2337280
4
  },
5
  "weight_map": {
6
  "model.embed_tokens.weight": "model-00001-of-00002.safetensors",
7
  "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
8
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
9
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
10
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
11
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
12
  "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
13
  "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
14
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
15
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
16
+ "model.layers.1.input_layernorm.weight": "model-00002-of-00002.safetensors",
17
+ "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
18
+ "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
19
+ "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
20
+ "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
21
+ "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
22
+ "model.layers.1.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
23
+ "model.layers.1.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
24
+ "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
25
+ "model.norm.weight": "model-00002-of-00002.safetensors",
26
+ "lm_head.weight": "model-00002-of-00002.safetensors"
27
  }
28
  }