asiansoul commited on
Commit
f92eb9d
·
verified ·
1 Parent(s): 810de9a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -0
README.md CHANGED
@@ -24,6 +24,7 @@ tags:
24
 
25
  "좋아(Joah)" by AsianSoul
26
 
 
27
  ## Merge Details
28
 
29
 
@@ -37,6 +38,13 @@ Don't worry even if you don't get the results you want.
37
 
38
  I'll find the answer for you.
39
 
 
 
 
 
 
 
 
40
  ### Merge Method
41
 
42
  This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.
 
24
 
25
  "좋아(Joah)" by AsianSoul
26
 
27
+
28
  ## Merge Details
29
 
30
 
 
38
 
39
  I'll find the answer for you.
40
 
41
+ Soon PoSE to extend Llama's context length to 64k with using my merge method : "reborn"[reborn](https://medium.com/@puffanddmx82/reborn-elevating-model-adaptation-with-merging-for-superior-nlp-performance-f604e8e307b2)
42
+
43
+ 256k is not possible. My computer is running out of memory. If you support me, i will try it on a computer with maximum specifications.
44
+
45
+ If you support me, i would like to conduct great tests by building a network with high-capacity traffic and high-speed 10G speeds for you.
46
+
47
+
48
  ### Merge Method
49
 
50
  This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base.