ITT-AF commited on
Commit
e3f18ff
1 Parent(s): 578655e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -1,3 +1,36 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+ ## ITT-AF/ITT-Yi-Ko-6B-v1.0
5
+ This model is a fine-tuned version of [beomi/Yi-Ko-6B](https://huggingface.co/beomi/Yi-Ko-6B) on an unknown dataset.
6
+
7
+ ### Model description
8
+ More information needed
9
+
10
+ ### Intended uses & limitations
11
+ More information needed
12
+
13
+ ### Training and evaluation data
14
+ More information needed
15
+
16
+ ### Training procedure
17
+ ### Training hypuerparameters
18
+ The following hyperparameters were used during training:
19
+ * learning_rate: 2e-05
20
+ * train_batch_size: 4
21
+ * eval_batch_size: 8
22
+ * seed: 42
23
+ * gradient_accumulation_steps: 8
24
+ * total_train_batch_size: 32
25
+ * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
26
+ * lr_scheduler_type: linear
27
+ * num_epochs: 1.0
28
+ * mixed_precision_training: Native AMP
29
+
30
+ ### Training results
31
+
32
+ ### Framework versions
33
+ * Transformers 4.36.2
34
+ * Pytorch 2.1.2+cu121
35
+ * Datasets 2.0.0
36
+ * Tokenizers 0.15.0