Text Generation
Transformers
bloom
Inference Endpoints
manojpreveen commited on
Commit
4f1cef7
·
1 Parent(s): 56460db

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +25 -0
README.md ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bigscience-openrail-m
3
+ datasets:
4
+ - manojpreveen/Instruction_Tuning
5
+ ---
6
+ Instruction Tuned Bloomz-7B1 model on ChatGPT dataset (85k data) using ***Colossal AI***
7
+
8
+ **Base Model:** bigscience/bloomz-7b1
9
+
10
+ **Training Details :**
11
+ * Epochs: 5
12
+ * Batch Size : 32 instantaneous per device x 1 gradient accumulation steps x 8 gpus = 256
13
+ * Max Length : 512
14
+ * Weight Decay : 0
15
+ * Learning Rate : 2e-5
16
+ * Learning Rate Scheduler Type : Cosine
17
+ * Number of warmup steps : 0
18
+ * Machine : 8xA100 80GB
19
+
20
+ **Dataset Details :**
21
+
22
+ Dataset : manojpreveen/Instruction_Tuning
23
+
24
+ Files :
25
+ * chat_gpt_v1.csv