File size: 2,805 Bytes
2bd25d9
 
 
 
 
 
 
 
 
f89a47d
2bd25d9
 
 
 
f89a47d
2bd25d9
98012a1
94ddbc1
2bd25d9
 
 
 
 
 
 
 
f89a47d
 
2bd25d9
 
 
082e2b5
 
2bd25d9
bbc9202
 
 
 
 
 
d5bd65d
c9b188e
bbc9202
2bd25d9
 
 
 
 
 
 
 
 
 
 
 
f89a47d
2bd25d9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
license: llama2
datasets:
- Open-Orca/OpenOrca
---
# Model Card for Model ID

<!-- Provide a quick summary of what the model is/does. -->

在llama-2-13b上使用open orca前20萬筆資料集進行訓練

# Fine-Tuning Information
- **GPU:** RTX4090 (single core / 24564MiB)
- **model:** meta-llama/Llama-2-13b-hf
- **dataset:** Open-Orca/OpenOrca (取前20w筆訓練集)
- **peft_type:** LoRA
- **lora_rank:** 8
- **lora_target:** q_proj, v_proj
- **per_device_train_batch_size:** 8
- **gradient_accumulation_steps:** 8
- **learning_rate :** 5e-5
- **epoch:** 1
- **precision:** bf16
- **quantization:** load_in_4bit

# Fine-Tuning Detail
- **train_loss:** 0.8616
- **train_runtime:** 29:18:07 (use deepspeed)

# Evaluation
- 評估結果來自**HuggingFaceH4/open_llm_leaderboard**
- 與Llama-2-13b和其他使用Open-Orca的模型比較4種Benchmark
- Benchmark包含**ARC****HellaSwag****MMLU****TruthfulQA**

| Model                                   |Average|  ARC  |HellaSwag| MMLU  | TruthfulQA |
|-----------------------------------------|-------|-------|---------|-------|------------|
|meta-llama/Llama-2-13b-hf                | 56.9  | 58.11 |  80.97  | 54.34 |   34.17    |
|meta-llama/Llama-2-13b-chat-hf           | 59.93 | 59.04 |  81.94  | 54.64 |   44.12    |
|Open-Orca/OpenOrca-Platypus2-13B         | 64.6  | 62.8  |  83.15  | 59.39 |   53.08    |
|Open-Orca/OpenOrcaxOpenChat-Preview2-13B | 63.81 | 62.37 |  82.96  | 58.68 |   51.23    |
|circulus/Llama-2-13b-orca-v1             | 62.91 | 62.03 |  82.27  | 57.71 |   49.61    |
|CHIH-HUNG/llama-2-13b-OpenOrca_5w        | 61.2  | 61.01 |  82.82  | 56.09 |   44.87    |
|CHIH-HUNG/llama-2-13b-open_orca_20w      | 60.46 | 59.9  |  82.51  | 56.3  |   43.14    |

# How to convert dataset to json

- 在**load_dataset**中輸入資料集名稱,並且在**take**中輸入要取前幾筆資料
- 觀察該資料集的欄位名稱,填入**example**欄位中(例如system_prompt、question、response)
- 最後指定json檔儲存位置 (**json_filename**)

```py
import json
from datasets import load_dataset

# 讀取數據集,take可以取得該數據集前n筆資料
dataset = load_dataset("Open-Orca/OpenOrca", split="train", streaming=True).take(200000)

# 提取所需欄位並建立新的字典列表
extracted_data = []
for example in dataset:
    extracted_example = {
        ### open orca
        "system_prompt": example["system_prompt"],
        "question": example["question"],
        "response": example["response"]
    }
    extracted_data.append(extracted_example)

# 指定 JSON 文件名稱
json_filename = "open_orca.json"

# 寫入 JSON 文件
with open(json_filename, "w") as json_file:
    json.dump(extracted_data, json_file, indent=4)

print(f"數據已提取並保存為 {json_filename}")
```