AndresR2909 commited on
Commit
7b9d41f
·
verified ·
1 Parent(s): c14d0f1

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -22,21 +22,6 @@
22
  "num_hidden_layers": 32,
23
  "num_key_value_heads": 8,
24
  "pretraining_tp": 1,
25
- "quantization_config": {
26
- "_load_in_4bit": true,
27
- "_load_in_8bit": false,
28
- "bnb_4bit_compute_dtype": "float16",
29
- "bnb_4bit_quant_storage": "uint8",
30
- "bnb_4bit_quant_type": "nf4",
31
- "bnb_4bit_use_double_quant": false,
32
- "llm_int8_enable_fp32_cpu_offload": false,
33
- "llm_int8_has_fp16_weight": false,
34
- "llm_int8_skip_modules": null,
35
- "llm_int8_threshold": 6.0,
36
- "load_in_4bit": true,
37
- "load_in_8bit": false,
38
- "quant_method": "bitsandbytes"
39
- },
40
  "rms_norm_eps": 1e-05,
41
  "rope_scaling": {
42
  "factor": 8.0,
@@ -47,7 +32,7 @@
47
  },
48
  "rope_theta": 500000.0,
49
  "tie_word_embeddings": false,
50
- "torch_dtype": "float32",
51
  "transformers_version": "4.44.2",
52
  "use_cache": true,
53
  "vocab_size": 128256
 
22
  "num_hidden_layers": 32,
23
  "num_key_value_heads": 8,
24
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  "rms_norm_eps": 1e-05,
26
  "rope_scaling": {
27
  "factor": 8.0,
 
32
  },
33
  "rope_theta": 500000.0,
34
  "tie_word_embeddings": false,
35
+ "torch_dtype": "float16",
36
  "transformers_version": "4.44.2",
37
  "use_cache": true,
38
  "vocab_size": 128256
generation_config.json CHANGED
@@ -6,7 +6,6 @@
6
  128008,
7
  128009
8
  ],
9
- "pad_token_id": 128009,
10
  "temperature": 0.6,
11
  "top_p": 0.9,
12
  "transformers_version": "4.44.2"
 
6
  128008,
7
  128009
8
  ],
 
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
  "transformers_version": "4.44.2"
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc913122b4a5230b43d796d5e26fadd669e1b9383d22916586b8401db735cfaa
3
+ size 4976698592
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c29485220c851efe964cc44eff4ea04d784bdfa7b112926be570b0f600ddf6e5
3
+ size 4999802616
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e8cb212179f57a1839fd177f6f4e2bd0ac2de8d2120cc7af56be2b790f89b560
3
+ size 4915916080
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e298ab0c60a5b9da77082fccb4e73eac9fd484688d3c95bd99616b81ce1ac86f
3
+ size 1168138808
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff