enable chatmode by default in KoboldAI

#1
by 11b - opened
config.json CHANGED
@@ -1,5 +1,6 @@
1
  {
2
  "_name_or_path": "pygmalion-6b",
 
3
  "activation_function": "gelu_new",
4
  "architectures": [
5
  "GPTJForCausalLM"
 
1
  {
2
  "_name_or_path": "pygmalion-6b",
3
+ "chatmode": true,
4
  "activation_function": "gelu_new",
5
  "architectures": [
6
  "GPTJForCausalLM"
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:978406b1338a4218387d4c9f6ca4ba5551077afc6c2dab2811aaeedc12b199b0
3
- size 10447397124
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45a3942e23a3a1ef354ad114367dfcc015296d7e0369db465e9762b13bd55740
3
+ size 10447065341
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dd67b8c3af2d341b347d11fd183fc9105ceb6b901b5339feaac5eafa1d908fd4
3
- size 5944343353
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05ff453b4b70604276b1c217c7e2479d07300bcadc23180867122e6fefc2dbc9
3
+ size 5530331025
tokenizer_config.json CHANGED
@@ -18,7 +18,7 @@
18
  },
19
  "errors": "replace",
20
  "model_max_length": 1024,
21
- "name_or_path": "pygmalion-6b",
22
  "special_tokens_map_file": null,
23
  "tokenizer_class": "GPT2Tokenizer",
24
  "unk_token": {
 
18
  },
19
  "errors": "replace",
20
  "model_max_length": 1024,
21
+ "name_or_path": "/home/amercurio/11b/ramdisk/convogpt/6b-uft",
22
  "special_tokens_map_file": null,
23
  "tokenizer_class": "GPT2Tokenizer",
24
  "unk_token": {