kyujinpy commited on
Commit
92a9614
1 Parent(s): e0bfeff

Upload 2 files

Browse files
Files changed (2) hide show
  1. PlatYi.png +0 -0
  2. README.md +66 -0
PlatYi.png ADDED
README.md CHANGED
@@ -1,3 +1,69 @@
1
  ---
 
 
 
 
 
 
2
  license: cc-by-nc-sa-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ datasets:
5
+ - garage-bAInd/Open-Platypus
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
  license: cc-by-nc-sa-4.0
9
  ---
10
+
11
+ # **PlatYi-34B-200k-Q-FastChat**
12
+ <img src='./PlatYi.png' width=256>
13
+
14
+ ## Model Details
15
+
16
+ **Model Developers** Kyujin Han (kyujinpy)
17
+
18
+ **Input** Models input text only.
19
+
20
+ **Output** Models generate text only.
21
+
22
+ **Model Architecture**
23
+ PlatYi-34B-200k-Q-FastChat is an auto-regressive language model based on the Yi-34B transformer architecture.
24
+
25
+ **Blog Link**
26
+ Blog: [Coming soon...]
27
+ Github: [Coming soon...]
28
+
29
+ **Base Model**
30
+ [01-ai/Yi-34B-200K](https://huggingface.co/01-ai/Yi-34B-200K)
31
+
32
+ **Training Dataset**
33
+ [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
34
+
35
+ **Notice**
36
+ While training, I used QLoRA.
37
+ `lora_r` values is 64.
38
+
39
+ **Apply prompting**
40
+ References by [FastChat](https://github.com/lm-sys/FastChat/blob/daa2b9abe20597ebf34dc5df164d450456610c74/fastchat/conversation.py#L198-L229).
41
+
42
+ # **Model Benchmark**
43
+
44
+ ## Open leaderboard
45
+ - Follow up as [link](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
46
+
47
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
48
+ | --- | --- | --- | --- | --- | --- | --- | --- |
49
+ | **PlatYi-34B-200k-Q-FastChat** | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
50
+ | PlatYi-34B-Llama-Q-FastChat | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
51
+ | [Yi-34B](https://huggingface.co/01-ai/Yi-34B) | 69.42 | 64.59 | 85.69 | 76.35 | 56.23 | 83.03 | 50.64 |
52
+
53
+ # Implementation Code
54
+ ```python
55
+ ### KO-Platypus
56
+ from transformers import AutoModelForCausalLM, AutoTokenizer
57
+ import torch
58
+
59
+ repo = "kyujinpy/PlatYi-34B-200k-Q-FastChat"
60
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
61
+ repo,
62
+ return_dict=True,
63
+ torch_dtype=torch.float16,
64
+ device_map='auto'
65
+ )
66
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
67
+ ```
68
+
69
+ ---