ITT-AF commited on
Commit
32a6737
1 Parent(s): e62ff8f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md CHANGED
@@ -1,3 +1,45 @@
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-4.0
3
  ---
4
+ # ITT-AF/ITT-42dot_LLM-PLM-1.3B-v1.0
5
+
6
+ This model is a fine-tuned version of [42dot/42dot_LLM-PLM-1.3B](https://huggingface.co/42dot/42dot_LLM-PLM-1.3B) on an custom dataset.
7
+
8
+ ## Model description
9
+
10
+ More information needed
11
+
12
+ ## Intended uses & limitations
13
+
14
+ More information needed
15
+
16
+ ## Training and evaluation data
17
+
18
+ More information needed
19
+
20
+ ## Training procedure
21
+
22
+ ### Training hyperparameters
23
+
24
+ The following hyperparameters were used during training:
25
+ - learning_rate: 2e-05
26
+ - train_batch_size: 24
27
+ - eval_batch_size: 8
28
+ - seed: 42
29
+ - gradient_accumulation_steps: 4
30
+ - total_train_batch_size: 96
31
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
32
+ - lr_scheduler_type: linear
33
+ - num_epochs: 1.0
34
+ - mixed_precision_training: Native AMP
35
+
36
+ ### Training results
37
+
38
+
39
+
40
+ ### Framework versions
41
+
42
+ - Transformers 4.36.2
43
+ - Pytorch 2.1.2+cu121
44
+ - Datasets 2.0.0
45
+ - Tokenizers 0.15.0