some1nostr commited on
Commit
f3848a8
1 Parent(s): 448db90

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -21
README.md CHANGED
@@ -8,57 +8,56 @@ license: apache-2.0
8
 
9
  # Model Card for Ostrich
10
 
 
11
 
12
- - **Trained with Nostr notes**
13
- - **Aligned a bit in these domains:**
14
- - Bitcoin
 
 
 
15
  - Health
16
  - Permaculture
17
  - Phytochemicals
18
  - Alternative medicine
19
  - Herbs
20
  - Nutrition
21
- - Faith
22
-
23
- You can read more about it here:
24
- https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xumrswfjx56rjwf4kkqhsx
25
-
26
-
27
 
28
- ## Model Details
29
 
30
-
31
- - **Finetuned from model:** https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct
32
-
33
- The number in the filenames like 7314 means the version.
34
 
35
 
36
  ## Uses
37
 
38
- Likes to disagree more compared to Llama3. It can be used for refutation, counter argument production.
39
 
40
- Ask any question, compared to other models this may know more about Nostr and Bitcoin.
41
-
42
- You can use llama.cpp to chat with it. You can also use llama-cpp-python package to use it in a Python script.
43
 
44
  Use repeat penalty of 1.05 or more to avoid repetitions.
45
 
 
 
46
 
47
  ## Warning
48
 
49
  Users (both direct and downstream) should be aware of the risks, biases and limitations of the model.
50
  The trainer, developer or uploader of this model does not assume any liability. Use it at your own risk.
51
 
52
-
53
  ## Training Details
54
 
55
  ### Training Data
56
 
57
- Nostr related info from web and nostr itself, bitcoin related info. Info on health domain.
58
  Information that aligns well with humanity is preferred.
 
59
 
60
  ### Training Procedure
61
 
62
- LLaMa-Factory is used to train on 2x3090! fsdp_qlora is the technique.
63
 
64
  The Nostr training took ~200 hours.
 
 
 
 
 
8
 
9
  # Model Card for Ostrich
10
 
11
+ **Finetuned from model:** [Llama3-70B](https://huggingface.co/meta-llama/Meta-Llama-3-70B-Instruct)
12
 
13
+ ## Fine Tuned With
14
+
15
+ **Trained with Nostr notes**: Knowledge and ideas about bitcoin and also Nostr itself. You can read more about the effects of Nostr training here:
16
+ https://habla.news/a/naddr1qvzqqqr4gupzp8lvwt2hnw42wu40nec7vw949ys4wgdvums0svs8yhktl8mhlpd3qqxnzde3xumrswfjx56rjwf4kkqhsx
17
+
18
+ **Aligned a bit in these domains:**
19
  - Health
20
  - Permaculture
21
  - Phytochemicals
22
  - Alternative medicine
23
  - Herbs
24
  - Nutrition
 
 
 
 
 
 
25
 
26
+ More to come around these topics.
27
 
28
+ Fine tuned a bit in faith related topics.
 
 
 
29
 
30
 
31
  ## Uses
32
 
33
+ Compared to other models this may know more about Nostr and Bitcoin. It is aligned with opinions of people on Nostr. It may have alternative ideas to mainstream because Nostr is very censorship resistant.
34
 
35
+ You can use llama.cpp to use the GGUF file.
 
 
36
 
37
  Use repeat penalty of 1.05 or more to avoid repetitions.
38
 
39
+ I am using the model here as a ground truth for Nostr related questions: https://wikifreedia.xyz/based-llm-leaderboard/npub1nlk894teh248w2heuu0x8z6jjg2hyxkwdc8cxgrjtm9lnamlskcsghjm9c
40
+
41
 
42
  ## Warning
43
 
44
  Users (both direct and downstream) should be aware of the risks, biases and limitations of the model.
45
  The trainer, developer or uploader of this model does not assume any liability. Use it at your own risk.
46
 
 
47
  ## Training Details
48
 
49
  ### Training Data
50
 
51
+ Nostr related info from web and nostr itself, bitcoin related info.
52
  Information that aligns well with humanity is preferred.
53
+ About 80% comes from Nostr notes. The rest is my own curation.
54
 
55
  ### Training Procedure
56
 
57
+ LLaMa-Factory is used to train on 2* RTX 3090! fsdp_qlora is the technique.
58
 
59
  The Nostr training took ~200 hours.
60
+
61
+ After training it for a while with notes, I used it to analyze and decide what notes to take in further for training.
62
+
63
+ The number in the filenames (like 9230) means the version. I take the training steps and add those to the version. There has been more than 9k training steps.