edgerunner-research commited on
Commit
621f7fe
1 Parent(s): d49c02b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -7,6 +7,9 @@ language:
7
 
8
  # EdgeRunner-Tactical-7B
9
 
 
 
 
10
  ## Introduction
11
 
12
  EdgeRunner-Tactical-7B is a powerful and efficient language model for the edge. Our mission is to build Generative AI for the edge that is safe, secure, and transparent. To that end, the EdgeRunner team is proud to release EdgeRunner-Tactical-7B, the most powerful language model for its size to date.
@@ -18,7 +21,7 @@ EdgeRunner-Tactical-7B is a 7 billion parameter language model that delivers pow
18
  - 7 billion parameters that balance power and efficiency
19
  - SOTA performance within the 7B model range
20
  - Initialized from Qwen2-Instruct, leveraging prior advancements
21
- - Self-Play Preference Optimization (SPPO) applied for continuous training and alignment
22
  - Competitive performance on several benchmarks with Meta’s Llama-3-70B, Mixtral 8x7B, and Yi 34B
23
  - Context length of 128K tokens, ideal for extensive conversations and large-scale text tasks
24
 
 
7
 
8
  # EdgeRunner-Tactical-7B
9
 
10
+
11
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/668ed3dcd857a9ca47edb75c/tSyuw39VtmEqvC_wptTDf.png)
12
+
13
  ## Introduction
14
 
15
  EdgeRunner-Tactical-7B is a powerful and efficient language model for the edge. Our mission is to build Generative AI for the edge that is safe, secure, and transparent. To that end, the EdgeRunner team is proud to release EdgeRunner-Tactical-7B, the most powerful language model for its size to date.
 
21
  - 7 billion parameters that balance power and efficiency
22
  - SOTA performance within the 7B model range
23
  - Initialized from Qwen2-Instruct, leveraging prior advancements
24
+ - [Self-Play Preference Optimization](https://arxiv.org/abs/2405.00675) (SPPO) applied for continuous training and alignment
25
  - Competitive performance on several benchmarks with Meta’s Llama-3-70B, Mixtral 8x7B, and Yi 34B
26
  - Context length of 128K tokens, ideal for extensive conversations and large-scale text tasks
27