xiaol commited on
Commit
4d509e3
1 Parent(s): 7bba5e2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +30 -2
README.md CHANGED
@@ -2,7 +2,35 @@
2
  license: apache-2.0
3
  ---
4
 
5
- More details: let me update lm eval harness to new version for more benchmarks.
6
- if you wanna help please contact me.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
 
8
  [Mobius-Chat-12B-128k](https://huggingface.co/TimeMobius/Mobius-Chat-12B-128k)
 
2
  license: apache-2.0
3
  ---
4
 
5
+ # Mobius Chat 12B 128K
6
+
7
+ ## Introduction
8
+
9
+ Mobius is a RWKV v5.2 arch model, a state based RNN+CNN+Transformer Mixed language model pretrained on a certain amount of data.
10
+ In comparison with the previous released Mobius, the improvements include:
11
+
12
+ * Only 24G Vram to run this model locally with fp16;
13
+ * Significant performance improvement;
14
+ * Multilingual support ;
15
+ * Stable support of 128K context length.
16
+
17
+
18
+ ## Usage
19
+ We encourage you use few shots to use this model, Desipte Directly use User: xxxx\n\nAssistant: xxx\n\n is really good too, Can boost all potential ability.
20
+
21
+ ## More details
22
+ Mobius 12B 128k based on RWKV v5.2 arch, which is leading state based RNN+CNN+Transformer Mixed language large language model which focus opensouce community
23
+ * 10~100 trainning/inference cost reduce;
24
+ * state based,which mean good at learning compression feature from language;
25
+ * community support.
26
+
27
+ ## requirements
28
+ 24G vram to run fp16, 12G for int8, 6G for nf4 with Ai00 server.
29
+
30
+ * [RWKV Runner](https://github.com/josStorer/RWKV-Runner)
31
+ * [Ai00 server](https://github.com/cgisky1980/ai00_rwkv_server)
32
+
33
+ ## future plan
34
+ If you need a HF version let us know
35
 
36
  [Mobius-Chat-12B-128k](https://huggingface.co/TimeMobius/Mobius-Chat-12B-128k)