uukuguy commited on
Commit
0b80597
1 Parent(s): 145e999

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -10
README.md CHANGED
@@ -34,23 +34,46 @@ The lora merged model is [here](https://huggingface.co/uukuguy/Mistral-7B-OpenOr
34
 
35
  The source code is [here](https://github.com/uukuguy/multi_loras)
36
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  ## lm-evaluation-harness
38
 
39
  [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
40
 
41
- | Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|
42
- | --- | --- |--- |
43
- | ARC | 64.08 | |
44
- | HellaSwag | 83.99 | |
45
- | MMLU | 62.24 | |
46
- | TruthfulQA | 53.05 | |
47
- | Average | 65.84 | |
48
 
49
  ## HumanEval
50
 
51
- | Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|
52
- | --- | --- | --- |
53
- | humaneval-python | 35.976 | |
54
 
55
 
56
  ## Training procedure
 
34
 
35
  The source code is [here](https://github.com/uukuguy/multi_loras)
36
 
37
+ ## Mistral-7B-OpenOrca
38
+
39
+ - Extract lora model [Mistral-7B-OpenOrca-lora](https://huggingface.co/uukuguy/Mistral-7B-OpenOrca-lora) from [Mistral-7B-OpenOrca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca);
40
+
41
+ - Merge the base model [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) with lora model to [Mistral-7B-OpenOrca-lora-merged](https://huggingface.co/uukuguy/Mistral-7B-OpenOrca-lora-merged)
42
+
43
+ - LLM Evaluation ...
44
+
45
+ ### Local Test
46
+
47
+ | | ARC_acc_norm (25-shot) | HellaSwag_acc_norm (10-shot) | MMLU_acc (5-shot) | TruthfulQA_mc2 (0-shot) | GSM8K_acc (8-shot) | Open LLM Score |
48
+ | ------ | ------ | ------ | ------ | ------ | ------ | ------ |
49
+ | Mistral-7B-OpenOrca | **71** | 83 | 61.42 | 45 | 40 | 65.11 |
50
+ | **r=64** | 67 | 84 | **64.26** | **47.32** | **41** | **65.65** |
51
+ | *r=16* | *65* | *83* | *62.84* | *46.95* | *38* | *64.45* |
52
+
53
+ ### Open LLM Leaderboard
54
+ | | ARC_acc_norm (25-shot) | HellaSwag_acc_norm (10-shot) | MMLU_acc (5-shot) | TruthfulQA_mc2 (0-shot) | Open LLM Score |
55
+ | ------ | ------ | ------ | ------ | ------ | ------ |
56
+ | Mistral-7B-SlimOrca | 62.54 | 83.86 | **62.77** | **54.23** | **65.85** |
57
+ | Mistral-7B-OpenOrca | **64.08** | **83.99** | 62.24 | 53.05 | 65.84 |
58
+
59
+
60
  ## lm-evaluation-harness
61
 
62
  [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
63
 
64
+ | Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora| Mistral-7B-OpenOrca-lora-merged |
65
+ | --- | --- |--- | --- |
66
+ | ARC | 64.08 | | |
67
+ | HellaSwag | 83.99 | | |
68
+ | MMLU | 62.24 | | |
69
+ | TruthfulQA | 53.05 | | |
70
+ | Average | 65.84 | | |
71
 
72
  ## HumanEval
73
 
74
+ | Metric | Mistral-7B-OpenOrca | Mistral-7B-OpenOrca-lora|Mistral-7B-OpenOrca-lora-merged |
75
+ | --- | --- | --- | --- |
76
+ | humaneval-python | 35.976 | | |
77
 
78
 
79
  ## Training procedure