totally-not-an-llm commited on
Commit
63bef5d
1 Parent(s): 795d32b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -0
README.md CHANGED
@@ -1,3 +1,41 @@
1
  ---
2
  license: llama2
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: llama2
3
+ datasets:
4
+ - totally-not-an-llm/EverythingLM-data-V2
5
  ---
6
+
7
+ # EverythingLM-13b-16k
8
+
9
+ Introducing EverythingLM, a llama-2 based, general-purpose 13b model with 16k context thanks to LlongMa. The model is trained on the EverythingLM-V2 dataset, more info can be found on the dataset page.
10
+
11
+ The model is completely uncensored.
12
+
13
+ ### GGML quants:
14
+ soon
15
+
16
+ Make sure to use correct rope scaling settings:
17
+ `-c 16384 --rope-freq-base 10000 --rope-freq-scale 0.25`
18
+ ### GPTQ quants:
19
+ soon
20
+
21
+ ### Notable features:
22
+ - Automatically triggered CoT reasoning.
23
+ - Verbose and detailed replies.
24
+ - Creative stories.
25
+ - Better prompt understanding.
26
+
27
+ ### Prompt format:
28
+ It is a modified Vicuna format, the same used in many of ehartford's models.
29
+ ```
30
+ You are a helpful AI assistant.
31
+
32
+ USER: <prompt>
33
+ ASSISTANT:
34
+ ```
35
+
36
+ Training took about 2.5 hours using QLoRa on 1xA100, so this model can be recreated for about $4. QLoRa model can be found here: https://huggingface.co/totally-not-an-llm/EverythingLM-13b-V2-peft.
37
+
38
+ ### Future plans:
39
+ - Native finetune.
40
+ - Other model sizes.
41
+ - Test some model merges using this model. (Specifically OpenOrca and Platypus models)