NeuralNovel
commited on
Commit
•
4a9eb91
1
Parent(s):
45dd19b
Update README.md
Browse files
README.md
CHANGED
@@ -10,16 +10,19 @@ tags:
|
|
10 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/645cfe4603fc86c46b3e46d1/1Uoxp_Bl9UwF-1K4KzwVN.png)
|
11 |
|
12 |
# ConvexAI/Solutus-3x7B
|
13 |
-
[Join our Discord!](https://discord.gg/XFChPkGd)
|
14 |
|
15 |
A model to test how MoE will route without square expansion.
|
16 |
|
|
|
|
|
17 |
The config looks like this...(detailed version is in the files and versions):
|
18 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
|
19 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
|
20 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
|
21 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
|
22 |
|
|
|
|
|
23 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
24 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
25 |
|
|
|
10 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/645cfe4603fc86c46b3e46d1/1Uoxp_Bl9UwF-1K4KzwVN.png)
|
11 |
|
12 |
# ConvexAI/Solutus-3x7B
|
|
|
13 |
|
14 |
A model to test how MoE will route without square expansion.
|
15 |
|
16 |
+
[Join our Discord!](https://discord.gg/rJXGjmxqzS)
|
17 |
+
|
18 |
The config looks like this...(detailed version is in the files and versions):
|
19 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - base
|
20 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #1
|
21 |
- [cognitivecomputations/WestLake-7B-v2-laser](https://huggingface.co/cognitivecomputations/WestLake-7B-v2-laser) - expert #2
|
22 |
- [mlabonne/Daredevil-7B](https://huggingface.co/mlabonne/Daredevil-7B) - expert #3
|
23 |
|
24 |
+
|
25 |
+
|
26 |
# "[What is a Mixture of Experts (MoE)?](https://huggingface.co/blog/moe)"
|
27 |
### (from the MistralAI papers...click the quoted question above to navigate to it directly.)
|
28 |
|