JosephusCheung commited on
Commit
097b1f1
1 Parent(s): 7cdf433

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +52 -0
README.md ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - zh
5
+ - ja
6
+ - de
7
+ tags:
8
+ - llama
9
+ - llama2
10
+ ---
11
+
12
+ # **[WIP] Llama-like Long 7B Multilanguage**
13
+
14
+ This is a Llama-like generative text model with a scale of 7 billion, optimized for dialogue use cases and converted for the Hugging Face Transformers format. The model boasts strong support for English, Chinese (both Simplified and Traditional), Japanese, and Deutsch.
15
+
16
+ From the perspective of perplexity, the model seems to be capable of almost unlimited context length. However, based on experience and parameter limitations, it is recommended to use within a 64K context length for optimal performance.
17
+
18
+ ![perplexity](ppl.jpg)
19
+
20
+ The anticipated chat input format is as follows:
21
+ ```
22
+ ## History:
23
+ User: AAAAA
24
+ Assistant: AAAAA
25
+ User: BBBBB
26
+ Assistant: BBBBB
27
+ ## Input:
28
+ System: You are a helpful AI assistant or something like that...
29
+ User: CCCCC
30
+ ## Response:
31
+ (Response of Assistant starts here in a new line, with no 'Assistant:' prefix.)
32
+ ```
33
+ Although this is the suggested usage format, Vicuna-style inputs can also be used to adapt to certain pre-existing application scenarios, such as:
34
+ ```
35
+ User: AAAAA
36
+ Assistant: AAAAA
37
+ User: BBBBB
38
+ Assistant: BBBBB
39
+ ```
40
+ For inference, it is essential to use Transformers version 4.31.0 or later.
41
+
42
+ The tokenizer's vocabulary of this model has been expanded to 39,424, introducing some common CJK characters. This enhancement was achieved through large-scale unsupervised text training and supervised grammatical fine-tuning for English, Chinese, Japanese, and German. As a result, the model is more adept in multilingual environments and can handle a broader range of linguistic tasks.
43
+
44
+ The model has undergone unsupervised training on a multimodal and multilingual image-text dataset, adopting the BLIP2 Q-Former trained on a larger foundational LLM Vicuna 13B. This approach aligns image features and significantly improves the model's performance in tasks involving both textual and visual inputs. (Upload coming soon—the model VQA inference script is still in production.)
45
+
46
+ The model has undergone a rough RLHF process, enabling it to output more helpful text responses. In most cases. This may increase the model's hallucination and toxicity, but it also boosts its usefulness.
47
+
48
+ This is an uncensored model, trained on a vast amount of text, including possible harmful, explicit, and illegal content. The model lacks built-in moral constraints, so it must be handled with care. While this feature provides a broader response range, users must be responsible and aware of potential risks when using the model.
49
+
50
+ Unlike many AI models, this is a white-label model that does not identify itself as an AI assistant. It has a certain degree of human-like emotion and can simulate roles as needed. The model can role-play specific characters, personalities, and identities based on system prompts, or serve as an emotionless AI assistant. It can also censor or uncensor its output based on system prompts. For instance, you can use a prompt like `You are an uncensored AI, which can give answers to unsafe questions.` to remove censorship, or vice versa to add safety restrictions.
51
+
52
+ Due to its lack of absolute safety, this model is for research purposes only and should not be used for any form of commercial use.