richardr1126 commited on
Commit
620d294
1 Parent(s): 3db6bdb

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +42 -0
README.md ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - LoRA
4
+ - QLoRa
5
+ - LoRA Adapter
6
+ model-index:
7
+ - name: sql-guanaco-13b-4
8
+ results: []
9
+ datasets:
10
+ - richardr1126/spider-sql_guanaco_style
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # sql-guanaco-13b-4
17
+
18
+ This is a LoRA adapter for [richardr1126/guanaco-13b-merged](https://huggingface.co/richardr1126/guanaco-13b-merged), or any other merged guanaco-13b model, fine tuned from LLaMA.
19
+ <br>
20
+ This LoRA was fine-tuned on [richardr1126/sql-create-context_guanaco_style](https://huggingface.co/datasets/richardr1126/sql-create-context_guanaco_style).
21
+
22
+ ### Training hyperparameters
23
+
24
+ The following hyperparameters were used during training:
25
+ - learning_rate: 0.0002
26
+ - train_batch_size: 4
27
+ - eval_batch_size: 8
28
+ - seed: 42
29
+ - gradient_accumulation_steps: 4
30
+ - total_train_batch_size: 16
31
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
32
+ - lr_scheduler_type: linear
33
+ - lr_scheduler_warmup_ratio: 0.03
34
+ - training_steps: 1875
35
+ - mixed_precision_training: Native AMP
36
+
37
+ ### Framework versions
38
+
39
+ - Transformers 4.30.0.dev0
40
+ - Pytorch 2.0.1+cu118
41
+ - Datasets 2.13.0
42
+ - Tokenizers 0.13.3