kyujinpy commited on
Commit
077ecb7
1 Parent(s): 38b83de

Upload 2 files

Browse files
Files changed (2) hide show
  1. PlatYi.png +0 -0
  2. README.md +66 -0
PlatYi.png ADDED
README.md CHANGED
@@ -1,3 +1,69 @@
1
  ---
 
 
 
 
 
 
2
  license: cc-by-nc-sa-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ language:
3
+ - en
4
+ datasets:
5
+ - garage-bAInd/Open-Platypus
6
+ library_name: transformers
7
+ pipeline_tag: text-generation
8
  license: cc-by-nc-sa-4.0
9
  ---
10
+
11
+ # **PlatYi-34B-Llama-Q**
12
+ <img src='./PlatYi.png' width=256>
13
+
14
+ ## Model Details
15
+
16
+ **Model Developers** Kyujin Han (kyujinpy)
17
+
18
+ **Input** Models input text only.
19
+
20
+ **Output** Models generate text only.
21
+
22
+ **Model Architecture**
23
+ PlatYi-34B-Llama-Q is an auto-regressive language model based on the Yi-34B transformer architecture.
24
+
25
+ **Blog Link**
26
+ Blog: [Coming soon...]
27
+ Github: [Coming soon...]
28
+
29
+ **Base Model**
30
+ [chargoddard/Yi-34B-Llama](https://huggingface.co/chargoddard/Yi-34B-Llama)
31
+
32
+ **Training Dataset**
33
+ [garage-bAInd/Open-Platypus](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
34
+
35
+ ## Notice
36
+ While training, I used Q-LoRA.
37
+ The lora_r values is 64.
38
+
39
+
40
+ # **Model Benchmark**
41
+
42
+ ## Open leaderboard
43
+ - Follow up as [link](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
44
+
45
+ | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
46
+ | --- | --- | --- | --- | --- | --- | --- | --- |
47
+ | **PlatYi-34B-Llama-Q** | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
48
+ | PlatYi-34B-Llama | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
49
+ | [Yi-34B-Llama](https://huggingface.co/chargoddard/Yi-34B-Llama) | 70.95 | 64.59 | 85.63 | 76.31 | 55.60 | 82.79 | 60.80 |
50
+ | [Yi-34B](https://huggingface.co/01-ai/Yi-34B) | 69.42 | 64.59 | 85.69 | 76.35 | 56.23 | 83.03 | 50.64 |
51
+
52
+
53
+ # Implementation Code
54
+ ```python
55
+ ### KO-Platypus
56
+ from transformers import AutoModelForCausalLM, AutoTokenizer
57
+ import torch
58
+
59
+ repo = "kyujinpy/PlatYi-34B-Llama-Q"
60
+ OpenOrca = AutoModelForCausalLM.from_pretrained(
61
+ repo,
62
+ return_dict=True,
63
+ torch_dtype=torch.float16,
64
+ device_map='auto'
65
+ )
66
+ OpenOrca_tokenizer = AutoTokenizer.from_pretrained(repo)
67
+ ```
68
+
69
+ ---