Text Generation
Transformers
GGUF
English
Inference Endpoints
leafspark commited on
Commit
5f7b78a
1 Parent(s): 90deb40

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -0
README.md CHANGED
@@ -1,3 +1,57 @@
1
  ---
2
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
+ datasets:
4
+ - HuggingFaceTB/cosmopedia
5
+ - databricks/databricks-dolly-15k
6
+ - Open-Orca/OpenOrca
7
+ language:
8
+ - en
9
+ metrics:
10
+ - accuracy
11
+ library_name: transformers
12
+ pipeline_tag: text-generation
13
  ---
14
+
15
+ # WikiChat-v0.2
16
+ Training in progress model to have conversations.
17
+
18
+ The GGUFs uploaded are full FP32 precision.
19
+
20
+ Using OpenOrca GPT-4 data + cosmopedia for some extra data + dolly15k for instruct
21
+
22
+ ## Model Details:
23
+ - 40M parameters
24
+ - 8 attention heads
25
+ - 32 layers
26
+ - 384 embeddings size
27
+ - 2048/8192 context (please use 4x RoPE scaling)
28
+
29
+ ## Prompt Format (Alpaca):
30
+ ```
31
+ Instruction: {system}
32
+ Input: {prompt}
33
+ Response: {response}
34
+ ```
35
+
36
+ Please structure your prompts in an instruct format for maximum performance.
37
+
38
+ ## Training Details:
39
+ - 1x RTX 3070 8GB (Infrencing speed: 80tok/s, full GPU offload)
40
+ - 1x Ryzen 3 3700x
41
+ - 96gb RAM
42
+ - 10 iterations
43
+ - Loss Target = 2.5 to 3.0
44
+ - Approx 30 samples (>0.0001 epoches)
45
+ - Training data = Refer to OpenOrca page
46
+
47
+ ## Notes:
48
+
49
+ The model isn't ready yet; this is to test tokenization of OpenOrca and a balance between training speed and model size
50
+
51
+ ## Example output:
52
+ ```
53
+ User: What is the square root of 4?
54
+ ```
55
+ ```
56
+ Assistant: The square root of 4 is 2.
57
+ ```