Lazycuber commited on
Commit
a648c51
1 Parent(s): dfd0f75

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "NousResearch/Llama-2-7b-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -21,6 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
  "transformers_version": "4.31.0",
24
- "use_cache": true,
25
  "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "NousResearch/Nous-Hermes-llama-2-7b",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
  "transformers_version": "4.31.0",
24
+ "use_cache": false,
25
  "vocab_size": 32000
26
  }
generation_config.json CHANGED
@@ -2,7 +2,7 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
- "pad_token_id": 32000,
6
  "temperature": 0.9,
7
  "top_p": 0.6,
8
  "transformers_version": "4.31.0"
 
2
  "_from_model_config": true,
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
+ "pad_token_id": 0,
6
  "temperature": 0.9,
7
  "top_p": 0.6,
8
  "transformers_version": "4.31.0"
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2019ca5255cdda8880be186cf6c6e1473a5c5f8372e4c9f1960c4d152b5c715
3
  size 9976638373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3efe363d2a902c19eadaeb08aaff42438cd2826facee010cbf16bb8287351c14
3
  size 9976638373
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e225cf1de06db75ff23af428d4981616031cc2530c51c7aede9c24cd4e93022
3
  size 3500317102
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12e32727e09bb0c3369350f5239ae4e81f89fff8681a734b9cc87c4da8cdb757
3
  size 3500317102