File size: 8,025 Bytes
022e112 33c9c02 5c0a394 33c9c02 5c0a394 022e112 5dfbc5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 |
---
language:
- en
license: cc-by-nc-4.0
library_name: transformers
tags:
- mergekit
- merge
base_model:
- SanjiWatsuki/Kunoichi-DPO-v2-7B
- macadeliccc/WestLake-7B-v2-laser-truthy-dpo
model-index:
- name: finch
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: AI2 Reasoning Challenge (25-Shot)
type: ai2_arc
config: ARC-Challenge
split: test
args:
num_few_shot: 25
metrics:
- type: acc_norm
value: 71.59
name: normalized accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: HellaSwag (10-Shot)
type: hellaswag
split: validation
args:
num_few_shot: 10
metrics:
- type: acc_norm
value: 87.87
name: normalized accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU (5-Shot)
type: cais/mmlu
config: all
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 64.81
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: TruthfulQA (0-shot)
type: truthful_qa
config: multiple_choice
split: validation
args:
num_few_shot: 0
metrics:
- type: mc2
value: 67.96
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: Winogrande (5-shot)
type: winogrande
config: winogrande_xl
split: validation
args:
num_few_shot: 5
metrics:
- type: acc
value: 84.14
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GSM8k (5-shot)
type: gsm8k
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 66.34
name: accuracy
source:
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=antiven0m/finch
name: Open LLM Leaderboard
---
<head> <link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/font-awesome/6.0.0-beta3/css/all.min.css"> </head> <style> body { font-family: "Helvetica Neue", Arial, sans-serif; background: radial-gradient(circle, #ffb347, #ffa92d, #ff9f14, #ff9500, #f08b00); color: #fff; line-height: 1.6; } .container { max-width: 800px; margin: 0 auto; padding: 40px; background-color: rgba(255, 255, 255, 0.1); border-radius: 10px; box-shadow: 0 0 20px rgba(0, 0, 0, 0.2); backdrop-filter: blur(10px); } .header { text-align: center; margin-bottom: 40px; } .title { font-size: 48px; font-weight: bold; text-transform: uppercase; letter-spacing: 2px; color: #fff; text-shadow: 2px 2px 4px rgba(0, 0, 0, 0.5); margin-bottom: 10px; } .subtitle { font-size: 24px; font-style: italic; color: #e6f7ff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); margin-bottom: 20px; } .gif { text-align: center; margin-bottom: 40px; } .gif img { max-width: 100%; height: auto; border-radius: 10px; box-shadow: 0 0 20px rgba(0, 0, 0, 0.3); } .info-section { margin-bottom: 40px; } .section-title { font-size: 32px; font-weight: bold; color: #e6f7ff; text-shadow: 2px 2px 4px rgba(0, 0, 0, 0.5); margin-bottom: 20px; position: relative; padding-left: 30px; } .section-title::before { content: ""; position: absolute; left: 0; top: 50%; transform: translateY(-50%); width: 20px; height: 20px; background-color: #e6f7ff; border-radius: 50%; box-shadow: 0 0 10px rgba(0, 0, 0, 0.3); } .info-item { background-color: rgba(255, 255, 255, 0.1); padding: 20px; border-radius: 10px; box-shadow: 0 0 10px rgba(0, 0, 0, 0.2); margin-bottom: 20px; } .info-item h3 { font-size: 24px; font-weight: bold; color: #e6f7ff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); margin-bottom: 10px; } .info-item p { font-size: 18px; color: #fff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); line-height: 1.4; } .info-item pre { background-color: rgba(0, 0, 0, 0.2); padding: 20px; border-radius: 10px; font-family: monospace; font-size: 16px; color: #fff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); overflow-x: auto; } .info-item a { color: #e6f7ff; text-decoration: none; border-bottom: 1px dashed #e6f7ff; transition: border-bottom 0.3s ease; } .info-item a:hover { border-bottom: 1px solid #e6f7ff; } .info-item table { width: 100%; border-collapse: collapse; box-shadow: 0 0 10px rgba(0, 0, 0, 0.2); } .info-item th, .info-item td { padding: 10px; text-align: left; border: 1px solid rgba(255, 255, 255, 0.2); } .info-item th { background-color: rgba(0, 0, 0, 0.2); font-weight: bold; color: #fff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); } .info-item td { color: #e6f7ff; text-shadow: 1px 1px 2px rgba(0, 0, 0, 0.3); } </style> <div class="container"> <div class="header"> <h1 class="title">Finch 7B Merge</h1> <p class="subtitle">A SLERP merge of two powerful 7B language models</p> </div> <div class="gif"> <img src="https://i.imgur.com/Da14544.gif" alt="Finch GIF"> </div> <div class="info-section"> <h2 class="section-title">Description</h2> <div class="info-item"> <p>Finch is a 7B language model created by merging <a href="https://huggingface.co/macadeliccc/WestLake-7B-v2-laser-truthy-dpo">macadeliccc/WestLake-7B-v2-laser-truthy-dpo</a> and <a href="https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B">SanjiWatsuki/Kunoichi-DPO-v2-7B</a> using the SLERP method.</p> </div> </div> <div class="info-section"> <h2 class="section-title">Quantized Models</h2> <div class="info-item"> <p>Quantized versions of Finch are available:</p> <ul> <li><a href="https://huggingface.co/antiven0m/finch-6bpw-exl2">6bpw EXL2 Quant</a></li> <li><a href="https://huggingface.co/antiven0m/finch-gguf">GGUF Quants</a></li> </ul> </div> </div> <div class="info-section"> <h2 class="section-title">Recommended Settings</h2> <div class="info-item"> <p>For best results, use the <b>ChatML</b> format with the following sampler settings:</p> <pre>Temperature: 1.2 Min P: 0.2 Smoothing Factor: 0.2</pre> </div> </div> <div class="info-section"> <h2 class="section-title">Mergekit Configuration</h2> <div class="info-item"> <pre>base_model: macadeliccc/WestLake-7B-v2-laser-truthy-dpo dtype: float16 merge_method: slerp parameters: t: - filter: self_attn value: [0.0, 0.5, 0.3, 0.7, 1.0] - filter: mlp value: [1.0, 0.5, 0.7, 0.3, 0.0] - value: 0.5 slices: - sources: - layer_range: [0, 32] model: macadeliccc/WestLake-7B-v2-laser-truthy-dpo - layer_range: [0, 32] model: SanjiWatsuki/Kunoichi-DPO-v2-7B</pre> </div> </div> <div class="info-section"> <h2 class="section-title">Evaluation Results</h2> <div class="info-item"> <p>Finch's performance on the Open LLM Leaderboard:</p> <table> <tr><th>Metric</th><th>Value</th></tr> <tr><td>Avg.</td><td>73.78</td></tr> <tr><td>AI2 Reasoning Challenge (25-Shot)</td><td>71.59</td></tr> <tr><td>HellaSwag (10-Shot)</td><td>87.87</td></tr> <tr><td>MMLU (5-Shot)</td><td>64.81</td></tr> <tr><td>TruthfulQA (0-shot)</td><td>67.96</td></tr> <tr><td>Winogrande (5-shot)</td><td>84.14</td></tr> <tr><td>GSM8k (5-shot)</td><td>66.34</td></tr> </table> <p>Detailed results: <a href="https://huggingface.co/datasets/open-llm-leaderboard/details_antiven0m__finch">https://huggingface.co/datasets/open-llm-leaderboard/details_antiven0m__finch</a></p> </div> </div> </div> |