Safetensors
Chinese
English
File size: 2,140 Bytes
5abc3d1
 
 
 
 
 
 
 
 
 
 
 
9b1d44f
5abc3d1
 
 
 
89420cf
 
 
df0a259
51bb0c9
44ab55f
89420cf
c245376
89420cf
 
44ab55f
 
 
 
 
 
 
 
 
 
 
 
c245376
 
 
9b1d44f
89420cf
 
 
9b1d44f
 
 
89420cf
 
 
 
 
 
44ab55f
89420cf
 
9b1d44f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: other
license_name: tongyi-qianwen
license_link: https://huggingface.co/Qwen/Qwen1.5-7B-Chat/blob/main/LICENSE
datasets:
- REILX/extracted_tagengo_gpt4
- TigerResearch/sft_zh
- alexl83/AlpacaDataCleaned
- LooksJuicy/ruozhiba
- silk-road/alpaca-data-gpt4-chinese
- databricks/databricks-dolly-15k
- microsoft/orca-math-word-problems-200k
- Sao10K/Claude-3-Opus-Instruct-5K
language:
- zh
- en
---


### 数据集
使用以下8个数据集
![image/png](https://cdn-uploads.huggingface.co/production/uploads/636f54b95d2050767e4a6317/OkuVQ1lWXRAKyel2Ef0Fz.png)
对Qwen1.5-7B-Chat进行微调并测试,结果显示,微调后的模型在CEVAL和MMLU的评分上均有所提升。

### 基础模型:
- https://huggingface.co/Qwen/Qwen1.5-7B-Chat


### 训练工具
https://github.com/hiyouga/LLaMA-Factory

### 测评方式:
使用opencompass(https://github.com/open-compass/OpenCompass/ ), 测试工具基于CEval和MMLU对微调之后的模型和原始模型进行测试。</br>
测试模型分别为:
- Qwen1.5-7B-Chat  
- Qwen1.5-7B-Chat-750Mb-lora,使用8DataSets数据集对Qwen1.5-7B-Chat模型进行sft方式lora微调

### 测试机器
8*A800

### 8DataSets数据集:
大约750Mb的微调数据集
- https://huggingface.co/datasets/REILX/extracted_tagengo_gpt4
- https://huggingface.co/datasets/TigerResearch/sft_zh
- https://huggingface.co/datasets/silk-road/alpaca-data-gpt4-chinese
- https://huggingface.co/datasets/LooksJuicy/ruozhiba
- https://huggingface.co/datasets/microsoft/orca-math-word-problems-200k
- https://huggingface.co/datasets/alexl83/AlpacaDataCleaned
- https://huggingface.co/datasets/Sao10K/Claude-3-Opus-Instruct-5K


### 结果
| 模型名称                 | CEVAL | MMLU |
|------------------------ |-------|------|
| Qwen1.5-7B-Chat         | 68.61 | 61.56 |
| Qwen1.5-7B-Chat-750Mb-lora  | 71.36 | 61.78 |

### License
This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses. The content of this project itself is licensed under the Apache license 2.0.