nRuaif commited on
Commit
7175c2e
1 Parent(s): a7ef563

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +86 -0
README.md ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: creativeml-openrail-m
3
+ language:
4
+ - en
5
+ pipeline_tag: text-generation
6
+ ---
7
+
8
+
9
+ ## Model Details
10
+
11
+ ### Model Description
12
+
13
+ <!-- Provide a longer summary of what this model is. -->
14
+
15
+
16
+
17
+ - **Developed by:** nRuaif
18
+ - **Model type:** large language model
19
+ - **License:**
20
+ - **Finetuned from model [optional]:** Llama-13B
21
+ ### Model Sources [optional]
22
+
23
+
24
+ ## Uses
25
+
26
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
27
+ The model uses Fastchat/ShareGPT format but anything would works fine
28
+
29
+
30
+ ### Direct Use
31
+
32
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
33
+
34
+ This model is finetuned for normal and erotic roleplay while can still an assistant. (Might not be a helpfull one through)
35
+
36
+
37
+
38
+ ### Out-of-Scope Use
39
+
40
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
41
+ Do anything you want. I don't care
42
+
43
+
44
+ ## Bias, Risks, and Limitations
45
+
46
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
47
+
48
+ Model might have bias to NSFW due to the large % of NSFW data in the training set.
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+ ## Training Details
57
+
58
+ ### Training Data
59
+
60
+ <!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
61
+
62
+
63
+ 3000 convos with 4090 cut off len.
64
+
65
+ ### Training Procedure
66
+
67
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
68
+
69
+
70
+
71
+
72
+ #### Training Hyperparameters
73
+
74
+ - **Training regime:** BF16, QLoRA, constant LR 5e-5 <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+ ### Compute Infrastructure
83
+
84
+ The model is trained on 1 A100 for 2 hours on runpod.
85
+
86
+