LeroyDyer commited on
Commit
f969e31
1 Parent(s): 08c87db

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +160 -5
README.md CHANGED
@@ -1,23 +1,178 @@
1
  ---
2
- base_model: LeroyDyer/_Spydaz_Web_AI_003
3
  language:
4
  - en
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
6
  tags:
7
  - text-generation-inference
8
  - transformers
9
  - unsloth
10
  - mistral
11
  - trl
12
- - sft
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
  ---
14
 
15
  # Uploaded model
16
 
17
- - **Developed by:** LeroyDyer
18
  - **License:** apache-2.0
19
- - **Finetuned from model :** LeroyDyer/_Spydaz_Web_AI_003
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
20
 
21
  This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
22
 
23
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
1
  ---
 
2
  language:
3
  - en
4
+ - sw
5
+ - ig
6
+ - so
7
+ - es
8
+ - ca
9
+ - xh
10
+ - zu
11
+ - ha
12
+ - tw
13
+ - af
14
+ - hi
15
+ - bm
16
+ - su
17
  license: apache-2.0
18
+ metrics:
19
+ - accuracy
20
+ - bertscore
21
+ - bleu
22
+ - brier_score
23
+ - cer
24
+ - character
25
+ - charcut_mt
26
+ - chrf
27
+ - code_eval
28
  tags:
29
  - text-generation-inference
30
  - transformers
31
  - unsloth
32
  - mistral
33
  - trl
34
+ - chemistry
35
+ - biology
36
+ - legal
37
+ - art
38
+ - music
39
+ - finance
40
+ - code
41
+ - medical
42
+ - not-for-all-audiences
43
+ - merge
44
+ - climate
45
+ - chain-of-thought
46
+ - tree-of-knowledge
47
+ - forest-of-thoughts
48
+ - visual-spacial-sketchpad
49
+ - alpha-mind
50
+ - knowledge-graph
51
+ - entity-detection
52
+ - encyclopedia
53
+ - wikipedia
54
+ - stack-exchange
55
+ - Reddit
56
+ - Cyber-series
57
+ - MegaMind
58
+ - Cybertron
59
+ - SpydazWeb
60
+ - Spydaz
61
+ - LCARS
62
+ - star-trek
63
+ - mega-transformers
64
+ - Mulit-Mega-Merge
65
+ - Multi-Lingual
66
+ - Afro-Centric
67
+ - African-Model
68
+ - Ancient-One
69
+ datasets:
70
+ - gretelai/synthetic_text_to_sql
71
+ - HuggingFaceTB/cosmopedia
72
+ - teknium/OpenHermes-2.5
73
+ - Open-Orca/SlimOrca
74
+ - Open-Orca/OpenOrca
75
+ - cognitivecomputations/dolphin-coder
76
+ - databricks/databricks-dolly-15k
77
+ - yahma/alpaca-cleaned
78
+ - uonlp/CulturaX
79
+ - mwitiderrick/SwahiliPlatypus
80
+ - swahili
81
+ - Rogendo/English-Swahili-Sentence-Pairs
82
+ - ise-uiuc/Magicoder-Evol-Instruct-110K
83
+ - meta-math/MetaMathQA
84
+ - abacusai/ARC_DPO_FewShot
85
+ - abacusai/MetaMath_DPO_FewShot
86
+ - abacusai/HellaSwag_DPO_FewShot
87
+ - HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset
88
+ - HuggingFaceFW/fineweb
89
+ - occiglot/occiglot-fineweb-v0.5
90
+ - omi-health/medical-dialogue-to-soap-summary
91
+ - keivalya/MedQuad-MedicalQnADataset
92
+ - ruslanmv/ai-medical-dataset
93
+ - Shekswess/medical_llama3_instruct_dataset_short
94
+ - ShenRuililin/MedicalQnA
95
+ - virattt/financial-qa-10K
96
+ - PatronusAI/financebench
97
+ - takala/financial_phrasebank
98
+ - Replete-AI/code_bagel
99
+ - athirdpath/DPO_Pairs-Roleplay-Alpaca-NSFW
100
+ - IlyaGusev/gpt_roleplay_realm
101
+ - rickRossie/bluemoon_roleplay_chat_data_300k_messages
102
+ - jtatman/hypnosis_dataset
103
+ - Hypersniper/philosophy_dialogue
104
+ - Locutusque/function-calling-chatml
105
+ - bible-nlp/biblenlp-corpus
106
+ - DatadudeDev/Bible
107
+ - Helsinki-NLP/bible_para
108
+ - HausaNLP/AfriSenti-Twitter
109
+ - aixsatoshi/Chat-with-cosmopedia
110
+ - HuggingFaceTB/cosmopedia-100k
111
+ - HuggingFaceFW/fineweb-edu
112
+ - m-a-p/CodeFeedback-Filtered-Instruction
113
+ - heliosbrahma/mental_health_chatbot_dataset
114
+ - saillab/alpaca_lingala_taco
115
+ - saillab/alpaca_igbo_taco
116
+ - saillab/alpaca_sundanese_taco
117
+ - saillab/alpaca_bambara_taco
118
+ - proj-persona/PersonaHub
119
+ - iamtarun/code_instructions_120k_alpaca
120
+ base_model: LeroyDyer/_Spydaz_Web_AI_004_bit
121
  ---
122
 
123
  # Uploaded model
124
 
125
+ - **Developed by:** Leroy "Spydaz" Dyer
126
  - **License:** apache-2.0
127
+ - **Finetuned from model :** LeroyDyer/LCARS_AI_009
128
+ [<img src="https://cdn-avatars.huggingface.co/v1/production/uploads/65d883893a52cd9bcd8ab7cf/tRsCJlHNZo1D02kBTmfy9.jpeg" width="300"/>
129
+ https://github.com/spydaz
130
+
131
+
132
+ * The Mistral-7B-Instruct-v0.2 Large Language Model (LLM) is an instruct fine-tuned version of the Mistral-7B-v0.2.
133
+
134
+ * Mistral-7B-v0.2 has the following changes compared to Mistral-7B-v0.1
135
+
136
+ * 32k context window (vs 8k context in v0.1)
137
+ * Rope-theta = 1e6
138
+ * No Sliding-Window Attention
139
+
140
+
141
+ # Introduction :
142
+
143
+ ## SpydazWeb AI model :
144
+
145
+ ### Methods:
146
+
147
+ Trained for multi-task operations as well as rag and function calling :
148
+
149
+ This model is a fully functioning model and is fully uncensored:
150
+
151
+ the model has been trained on multiple datasets on the huggingface hub and kaggle :
152
+
153
+ the focus has been mainly on methodology :
154
+
155
+ * Chain of thoughts
156
+ * steo by step
157
+ * tree of thoughts
158
+ * forest of thoughts
159
+ * graph of thoughts
160
+ * agent generation : Voting, ranking, ...
161
+
162
+ with these methods the model has gained insights into tasks, enabling for knowldge transfer between tasks :
163
+
164
+ the model has been intensivly trained in recalling data previously entered into the matrix:
165
+
166
+
167
+
168
+
169
+
170
+
171
+
172
+
173
+
174
+
175
 
176
  This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
177
 
178
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)