edgerunner-research
commited on
Commit
•
621f7fe
1
Parent(s):
d49c02b
Update README.md
Browse files
README.md
CHANGED
@@ -7,6 +7,9 @@ language:
|
|
7 |
|
8 |
# EdgeRunner-Tactical-7B
|
9 |
|
|
|
|
|
|
|
10 |
## Introduction
|
11 |
|
12 |
EdgeRunner-Tactical-7B is a powerful and efficient language model for the edge. Our mission is to build Generative AI for the edge that is safe, secure, and transparent. To that end, the EdgeRunner team is proud to release EdgeRunner-Tactical-7B, the most powerful language model for its size to date.
|
@@ -18,7 +21,7 @@ EdgeRunner-Tactical-7B is a 7 billion parameter language model that delivers pow
|
|
18 |
- 7 billion parameters that balance power and efficiency
|
19 |
- SOTA performance within the 7B model range
|
20 |
- Initialized from Qwen2-Instruct, leveraging prior advancements
|
21 |
-
- Self-Play Preference Optimization (SPPO) applied for continuous training and alignment
|
22 |
- Competitive performance on several benchmarks with Meta’s Llama-3-70B, Mixtral 8x7B, and Yi 34B
|
23 |
- Context length of 128K tokens, ideal for extensive conversations and large-scale text tasks
|
24 |
|
|
|
7 |
|
8 |
# EdgeRunner-Tactical-7B
|
9 |
|
10 |
+
|
11 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/668ed3dcd857a9ca47edb75c/tSyuw39VtmEqvC_wptTDf.png)
|
12 |
+
|
13 |
## Introduction
|
14 |
|
15 |
EdgeRunner-Tactical-7B is a powerful and efficient language model for the edge. Our mission is to build Generative AI for the edge that is safe, secure, and transparent. To that end, the EdgeRunner team is proud to release EdgeRunner-Tactical-7B, the most powerful language model for its size to date.
|
|
|
21 |
- 7 billion parameters that balance power and efficiency
|
22 |
- SOTA performance within the 7B model range
|
23 |
- Initialized from Qwen2-Instruct, leveraging prior advancements
|
24 |
+
- [Self-Play Preference Optimization](https://arxiv.org/abs/2405.00675) (SPPO) applied for continuous training and alignment
|
25 |
- Competitive performance on several benchmarks with Meta’s Llama-3-70B, Mixtral 8x7B, and Yi 34B
|
26 |
- Context length of 128K tokens, ideal for extensive conversations and large-scale text tasks
|
27 |
|