Update README.md
Browse files
README.md
CHANGED
@@ -24,6 +24,7 @@ tags:
|
|
24 |
|
25 |
"좋아(Joah)" by AsianSoul
|
26 |
|
|
|
27 |
## Merge Details
|
28 |
|
29 |
|
@@ -37,6 +38,13 @@ Don't worry even if you don't get the results you want.
|
|
37 |
|
38 |
I'll find the answer for you.
|
39 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
40 |
### Merge Method
|
41 |
|
42 |
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.
|
|
|
24 |
|
25 |
"좋아(Joah)" by AsianSoul
|
26 |
|
27 |
+
|
28 |
## Merge Details
|
29 |
|
30 |
|
|
|
38 |
|
39 |
I'll find the answer for you.
|
40 |
|
41 |
+
Soon PoSE to extend Llama's context length to 64k with using my merge method : "reborn"[reborn](https://medium.com/@puffanddmx82/reborn-elevating-model-adaptation-with-merging-for-superior-nlp-performance-f604e8e307b2)
|
42 |
+
|
43 |
+
256k is not possible. My computer is running out of memory. If you support me, i will try it on a computer with maximum specifications.
|
44 |
+
|
45 |
+
If you support me, i would like to conduct great tests by building a network with high-capacity traffic and high-speed 10G speeds for you.
|
46 |
+
|
47 |
+
|
48 |
### Merge Method
|
49 |
|
50 |
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.
|