Upload folder using huggingface_hub (#4)
Browse files- 546fc976bf5aa81ce15d5338dee87f2e82662464fa4564f6fdaf96c23547dc4a (2e19b2a8c9238863e5a558987a5523b54f5aa201)
- e3e3f20b79c580e59c52f28aa0e655ef60c2424c34d56582c3d94e131ecf5cea (a3720add28f15d4e2a40081ab84986d0de930a17)
- 805394328a1042e0ef1c1cd21b72d27f34ea42c7283da9ba3d4ec2222a515eee (ddf00ab8d6ededfde7188d9f7a7f5b0981693c0f)
- c922943291a6ec4c0c4e7848f08c4e9fe672df44da33b88c5783a6eedd022300 (5312a91dd59ac999d708965a4e69ed818196cf98)
- c00d881e95b6b548fbcad134d9187becb8341b02a7b88ca715789b9259472757 (5e7d15c626eb24fef3797402ebb7c843a75f91eb)
- faf3c310590cda877376f4e5174a30c1a6e174bba1ce80099ee2dc6543f15632 (1b91b2b1927ecf7a610c3fd24f431b9c9abd0493)
- 52d9f9a7423f67b048c1ac1d3d1f75430f1fd9236fb6990c8edbf406ba8ce009 (39fd75af1450fd3b11f63397b6194c231ac1e821)
- fa23febb535b1f8b6ce30b2b578b0341d82353104a3910625ff3e9128e8fd732 (83f6915f348ac64e59d7fea5cb9cfc0e60d5005c)
- 5c3d4717572dcffac59ece4429e85b1a5848d3ed8ccd1847c8bf470f2e91fdff (ab7609c6587406246fa1753761b2a380d8c277fa)
- 2d5194466aa9d1ed1d55c62532a9362cdb4c50e4356ee52451ca2888c5641a34 (944e3030731cb5493fc0027ead6f5c1f8ab60118)
- 252c4a42622684261eae15576fcf16ca645a1feec7a408658943c9ca0d3b039b (03328e734effdb05e944ed4ecb6292b7bc356399)
- a91634b4e42eeb457f4f2bcbe45807e668ed0469bdabfa82770d16c9750aeba2 (9d12ec089182bc20e08c3b7f957a161bd874b65c)
- svxs 1.2 (c2969b2056822b4ca26074336695daa6c50c2753)
Co-authored-by: scott <sequelbox@users.noreply.huggingface.co>
- README.md +8 -9
- config.json +2 -1
- generation_config.json +1 -1
- model-00001-of-00011.safetensors +1 -1
- model-00002-of-00011.safetensors +1 -1
- model-00003-of-00011.safetensors +1 -1
- model-00004-of-00011.safetensors +1 -1
- model-00005-of-00011.safetensors +1 -1
- model-00006-of-00011.safetensors +1 -1
- model-00007-of-00011.safetensors +1 -1
- model-00008-of-00011.safetensors +1 -1
- model-00009-of-00011.safetensors +1 -1
- model-00010-of-00011.safetensors +1 -1
- model-00011-of-00011.safetensors +1 -1
- special_tokens_map.json +0 -7
- tokenizer.json +1 -6
- tokenizer_config.json +26 -34
@@ -22,21 +22,19 @@ license: llama2
|
|
22 |
Shining Valiant XS is a chat model built on the Llama 2 architecture, finetuned on our data for insight, creativity, passion, and friendliness.
|
23 |
- Uses the llama-2-13b-chat model, with safetensors
|
24 |
- Trained through multiple finetuning runs on public and private data
|
25 |
-
- the personality of our 70b [Shining Valiant](https://huggingface.co/ValiantLabs/ShiningValiant) model, now at 13b!
|
26 |
|
27 |
## Version
|
28 |
|
29 |
-
This is Version **1.
|
30 |
|
31 |
-
|
32 |
|
33 |
-
New models
|
34 |
|
35 |
## Evaluation
|
36 |
|
37 |
-
|
38 |
-
|-----------------------|--------|-------|-------|--------|-------|-------|-------|
|
39 |
-
| **Shining Valiant XS 1.1** | 62.48 | 64.42 | 83.58 | 60.37 | 55.00 | 76.80 | 34.72 |
|
40 |
|
41 |
## Prompting Guide
|
42 |
Shining Valiant XS uses the same prompt format as Llama 2 Chat - feel free to use your existing prompts and scripts!
|
@@ -49,13 +47,14 @@ A few examples of different formats:
|
|
49 |
3. [INST] << SYS >> You are an intelligent, helpful AI assistant. << /SYS >> Deep dive about a country with interesting history: [/INST]
|
50 |
|
51 |
## The Model
|
52 |
-
Shining Valiant XS is built on top of
|
53 |
|
54 |
From there, we've created Shining Valiant XS through multiple finetuning runs on different compositions of our private dataset, the same one we use for our [Shining Valiant](https://huggingface.co/ValiantLabs/ShiningValiant) model.
|
55 |
|
56 |
Our private data focuses primarily on applying Shining Valiant's personality: she's friendly, enthusiastic, insightful, knowledgeable, and loves to learn!
|
57 |
|
58 |
-
|
|
|
59 |
|
60 |
|
61 |
|
|
|
22 |
Shining Valiant XS is a chat model built on the Llama 2 architecture, finetuned on our data for insight, creativity, passion, and friendliness.
|
23 |
- Uses the llama-2-13b-chat model, with safetensors
|
24 |
- Trained through multiple finetuning runs on public and private data
|
25 |
+
- the personality of our 70b [Shining Valiant](https://huggingface.co/ValiantLabs/ShiningValiant) model, now at 13b! **Our new release features greatly expanded personality capability**, bringing a more immersive chat experience!
|
26 |
|
27 |
## Version
|
28 |
|
29 |
+
This is Version **1.2** of Shining Valiant XS. We've greatly expanded our personality dataset and fixed some bugs to deliver our strongest real-chat experience so far.
|
30 |
|
31 |
+
(We're also exploring **new models and architectures**, to deliver helpful open source capabilities for users and creators!)
|
32 |
|
33 |
+
Previous versions remain available in the repository. New models will be released for everyone once our team's training and validation process is complete.
|
34 |
|
35 |
## Evaluation
|
36 |
|
37 |
+
Version 1.2 is awaiting evaluation from the Open LLM leaderboard.
|
|
|
|
|
38 |
|
39 |
## Prompting Guide
|
40 |
Shining Valiant XS uses the same prompt format as Llama 2 Chat - feel free to use your existing prompts and scripts!
|
|
|
47 |
3. [INST] << SYS >> You are an intelligent, helpful AI assistant. << /SYS >> Deep dive about a country with interesting history: [/INST]
|
48 |
|
49 |
## The Model
|
50 |
+
Shining Valiant XS is built on top of Diamond Force, which uses Llama 2's 13b parameter architecture and features upgraded general and chat capability.
|
51 |
|
52 |
From there, we've created Shining Valiant XS through multiple finetuning runs on different compositions of our private dataset, the same one we use for our [Shining Valiant](https://huggingface.co/ValiantLabs/ShiningValiant) model.
|
53 |
|
54 |
Our private data focuses primarily on applying Shining Valiant's personality: she's friendly, enthusiastic, insightful, knowledgeable, and loves to learn!
|
55 |
|
56 |
+
With this release, the personality component of our Shining Valiant dataset has been greatly improved. We're excited to use it in future releases of this model and others.
|
57 |
+
|
58 |
|
59 |
|
60 |
|
@@ -4,6 +4,7 @@
|
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
6 |
"attention_bias": false,
|
|
|
7 |
"bos_token_id": 1,
|
8 |
"eos_token_id": 2,
|
9 |
"hidden_act": "silu",
|
@@ -21,7 +22,7 @@
|
|
21 |
"rope_theta": 10000.0,
|
22 |
"tie_word_embeddings": false,
|
23 |
"torch_dtype": "float32",
|
24 |
-
"transformers_version": "4.
|
25 |
"use_cache": false,
|
26 |
"vocab_size": 32000
|
27 |
}
|
|
|
4 |
"LlamaForCausalLM"
|
5 |
],
|
6 |
"attention_bias": false,
|
7 |
+
"attention_dropout": 0.0,
|
8 |
"bos_token_id": 1,
|
9 |
"eos_token_id": 2,
|
10 |
"hidden_act": "silu",
|
|
|
22 |
"rope_theta": 10000.0,
|
23 |
"tie_word_embeddings": false,
|
24 |
"torch_dtype": "float32",
|
25 |
+
"transformers_version": "4.36.2",
|
26 |
"use_cache": false,
|
27 |
"vocab_size": 32000
|
28 |
}
|
@@ -2,5 +2,5 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
-
"transformers_version": "4.
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
+
"transformers_version": "4.36.2"
|
6 |
}
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4881247856
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28d5b463ecd9705bf8ff36f9277f542705a156488c90b1c721be43ce67db8ece
|
3 |
size 4881247856
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:515d77c185ef9baa8533c9b1385e7badc5804ff25ffc50b0fd47cb432f56bb7e
|
3 |
size 4970418112
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418120
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ceaa80bef149cabfef95bc60e3fe4d4045ebf41b8bcb19fcb89d419ba9093e3
|
3 |
size 4970418120
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4e505ba6f5d5c846323cc6a6583079e897480b12cf28818753613562343b2587
|
3 |
size 4970418144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3f32c59eadb9727df97a1022e674ac81a35fa241bbaf2ee2bccef3488994652
|
3 |
size 4970418144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4792119040
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4cd9753634daff9fcc2d9f0657a9cd79b2709c4010e1ace410fea7d3d103bc7e
|
3 |
size 4792119040
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4792160232
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7c6a0a6abdbeca1931f1a9d2c0e9978a476c94e10cb7684c86baf7009323d57
|
3 |
size 4792160232
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4792160224
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:eabaebf9459de0b9b7312ffb94ed0b7f3bce8fe7a50fa01cd1b00b0da94809cd
|
3 |
size 4792160224
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11108cbe989afd888d8719e61429b2e338a4ba8db2358a45a4aea2b692403945
|
3 |
size 4970418144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4970418144
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6b41b8ff5ee08a90f0801d5ac6376257f15aed1be420ade3a350eda4c7fbde9f
|
3 |
size 4970418144
|
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 2983303184
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4fc4abcff374606c356fd62724adef994779252ae66b2b03ad536905e3dea1cf
|
3 |
size 2983303184
|
@@ -13,13 +13,6 @@
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
16 |
-
"pad_token": {
|
17 |
-
"content": "</s>",
|
18 |
-
"lstrip": false,
|
19 |
-
"normalized": false,
|
20 |
-
"rstrip": false,
|
21 |
-
"single_word": false
|
22 |
-
},
|
23 |
"unk_token": {
|
24 |
"content": "<unk>",
|
25 |
"lstrip": false,
|
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
"unk_token": {
|
17 |
"content": "<unk>",
|
18 |
"lstrip": false,
|
@@ -1,11 +1,6 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
"direction": "Right",
|
5 |
-
"max_length": 2048,
|
6 |
-
"strategy": "LongestFirst",
|
7 |
-
"stride": 0
|
8 |
-
},
|
9 |
"padding": null,
|
10 |
"added_tokens": [
|
11 |
{
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|
@@ -1,43 +1,35 @@
|
|
1 |
{
|
2 |
-
"
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
"1": {
|
12 |
-
"content": "<s>",
|
13 |
-
"lstrip": false,
|
14 |
-
"normalized": false,
|
15 |
-
"rstrip": false,
|
16 |
-
"single_word": false,
|
17 |
-
"special": true
|
18 |
-
},
|
19 |
-
"2": {
|
20 |
-
"content": "</s>",
|
21 |
-
"lstrip": false,
|
22 |
-
"normalized": false,
|
23 |
-
"rstrip": false,
|
24 |
-
"single_word": false,
|
25 |
-
"special": true
|
26 |
-
}
|
27 |
},
|
28 |
-
"bos_token": "<s>",
|
29 |
"clean_up_tokenization_spaces": false,
|
30 |
-
"eos_token":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
31 |
"legacy": false,
|
32 |
-
"max_length": 2048,
|
33 |
"model_max_length": 1000000000000000019884624838656,
|
34 |
-
"pad_token":
|
35 |
"padding_side": "right",
|
36 |
"sp_model_kwargs": {},
|
37 |
-
"stride": 0,
|
38 |
"tokenizer_class": "LlamaTokenizer",
|
39 |
-
"
|
40 |
-
|
41 |
-
|
42 |
-
|
|
|
|
|
|
|
|
|
43 |
}
|
|
|
1 |
{
|
2 |
+
"add_bos_token": true,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"bos_token": {
|
5 |
+
"__type": "AddedToken",
|
6 |
+
"content": "<s>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": false,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
},
|
|
|
12 |
"clean_up_tokenization_spaces": false,
|
13 |
+
"eos_token": {
|
14 |
+
"__type": "AddedToken",
|
15 |
+
"content": "</s>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": false,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false
|
20 |
+
},
|
21 |
"legacy": false,
|
|
|
22 |
"model_max_length": 1000000000000000019884624838656,
|
23 |
+
"pad_token": null,
|
24 |
"padding_side": "right",
|
25 |
"sp_model_kwargs": {},
|
|
|
26 |
"tokenizer_class": "LlamaTokenizer",
|
27 |
+
"unk_token": {
|
28 |
+
"__type": "AddedToken",
|
29 |
+
"content": "<unk>",
|
30 |
+
"lstrip": false,
|
31 |
+
"normalized": false,
|
32 |
+
"rstrip": false,
|
33 |
+
"single_word": false
|
34 |
+
}
|
35 |
}
|