some1nostr commited on
Commit
138ba60
·
verified ·
1 Parent(s): 66e3803

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -10,8 +10,8 @@ license: apache-2.0
10
 
11
 
12
  - **Trained with some of the Nostr notes**
13
- - **Likes to disagree more compared to Llama3. It can be used for refutation, counter argument production.**
14
  - **Aligned a bit in these domains:**
 
15
  - Health
16
  - Permaculture
17
  - Phytochemicals
@@ -22,6 +22,10 @@ license: apache-2.0
22
  Read more about it here:
23
  https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xsunjwfkxcunwv3jvtnjyc
24
 
 
 
 
 
25
  ## Model Details
26
 
27
 
@@ -32,11 +36,11 @@ The number in the filenames like 4750 means the version. Higher numbers are newe
32
 
33
  ## Uses
34
 
 
 
35
  Ask any question, compared to other models this may know more about Nostr and Bitcoin.
36
- You can use llama.cpp to chat with it.
37
- You can also use llama-cpp-python package to use it in a Python script.
38
 
39
- Llama3 chat template can be used. <|begin_of_text|><|start_header_id|> ...
40
 
41
  Use repeat penalty of 1.05 or more to avoid repetitions.
42
 
@@ -58,4 +62,4 @@ Information that aligns well with humanity is preferred.
58
 
59
  LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
60
 
61
- The Nostr training took ~30 hours for a dataset of about 20MB.
 
10
 
11
 
12
  - **Trained with some of the Nostr notes**
 
13
  - **Aligned a bit in these domains:**
14
+ - Bitcoin
15
  - Health
16
  - Permaculture
17
  - Phytochemicals
 
22
  Read more about it here:
23
  https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xsunjwfkxcunwv3jvtnjyc
24
 
25
+ A running model is here:
26
+ https://njump.me/npub1chadadwep45t4l7xx9z45p72xsxv7833zyy4tctdgh44lpc50nvsrjex2m (Though it may be down for maintenance time to time. You need to DM the bot for it to answer.)
27
+
28
+
29
  ## Model Details
30
 
31
 
 
36
 
37
  ## Uses
38
 
39
+ Likes to disagree more compared to Llama3. It can be used for refutation, counter argument production.
40
+
41
  Ask any question, compared to other models this may know more about Nostr and Bitcoin.
 
 
42
 
43
+ You can use llama.cpp to chat with it. You can also use llama-cpp-python package to use it in a Python script.
44
 
45
  Use repeat penalty of 1.05 or more to avoid repetitions.
46
 
 
62
 
63
  LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
64
 
65
+ The Nostr training took ~140 hours for a dataset of ~67MB.