Knightcodin commited on
Commit
37d4d58
1 Parent(s): 952ed4f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -16
README.md CHANGED
@@ -1,20 +1,20 @@
1
  ---
 
 
2
  language:
3
  - en
4
- base_model: meta-llama/Meta-Llama-3-8B-Instruct
5
  pipeline_tag: text-generation
6
- tags:
7
- - facebook
8
- - meta
9
- - pytorch
10
- - llama
11
- - llama-3
12
- - moe
13
- - code
14
- license: other
15
- license_name: llama3
16
- license_link: LICENSE
17
  ---
 
 
 
 
 
 
 
 
 
18
 
19
 
20
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/657eb5b256c9c67605a6e8b5/8JXktjAyUPCWQGnRExiVI.png)
@@ -23,10 +23,6 @@ license_link: LICENSE
23
 
24
  Aplite-Instruct-4x8B-Llama-3 is a experimental MoE (Mixture of Experts) model based on the Llama-3 architecture using Mergekit.
25
 
26
- **GGUF Quant [Here](https://huggingface.co/raincandy-u/Llama-3-Aplite-Instruct-4x8B-GGUF)**
27
-
28
- **<|eot_id|> bug has been fixed.**
29
-
30
  ## Disclaimer
31
 
32
  This model is a research experiment and may generate incorrect or harmful content. The model's outputs should not be taken as factual or representative of the views of the model's creator or any other individual.
 
1
  ---
2
+ ---
3
+ license: cc-by-nc-4.0
4
  language:
5
  - en
6
+ quantized_by: KnightCodin
7
  pipeline_tag: text-generation
 
 
 
 
 
 
 
 
 
 
 
8
  ---
9
+ ## Exllama v2 Quantizations of L3-Solana-8B-v1
10
+
11
+ Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.19">turboderp's ExLlamaV2 v0.0.19</a> for quantization.
12
+
13
+ <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b>
14
+
15
+ Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions.
16
+
17
+ Original model: https://huggingface.co/raincandy-u/Llama-3-Aplite-Instruct-4x8B
18
 
19
 
20
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/657eb5b256c9c67605a6e8b5/8JXktjAyUPCWQGnRExiVI.png)
 
23
 
24
  Aplite-Instruct-4x8B-Llama-3 is a experimental MoE (Mixture of Experts) model based on the Llama-3 architecture using Mergekit.
25
 
 
 
 
 
26
  ## Disclaimer
27
 
28
  This model is a research experiment and may generate incorrect or harmful content. The model's outputs should not be taken as factual or representative of the views of the model's creator or any other individual.