noom83 commited on
Commit
22e6b89
·
verified ·
1 Parent(s): 7bd4205

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. README.md +35 -0
  2. model.pth +3 -0
  3. training_config.json +12 -0
README.md ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # aaa-2-sql
2
+
3
+ This is a finetuned version of [Mistral-7B-Instruct-v0.3](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3) using LoRA with LitGPT.
4
+
5
+ ## Training Details
6
+
7
+ - **Base Model:** mistralai/Mistral-7B-Instruct-v0.3
8
+ - **Framework:** LitGPT
9
+ - **Finetuning Method:** Low-Rank Adaptation (LoRA)
10
+ - **LoRA Parameters:**
11
+ - Rank (r): 16
12
+ - Alpha: 32
13
+ - Dropout: 0.05
14
+ - **Quantization:** bnb.nf4
15
+ - **Context Length:** 4098 tokens
16
+ - **Training Steps:** 2000
17
+
18
+ ## Usage
19
+
20
+ ```python
21
+ from transformers import AutoTokenizer, AutoModelForCausalLM
22
+
23
+ # Load model and tokenizer
24
+ model = AutoModelForCausalLM.from_pretrained("exaler/aaa-2-sql")
25
+ tokenizer = AutoTokenizer.from_pretrained("exaler/aaa-2-sql")
26
+
27
+ # Create prompt
28
+ prompt = "Your prompt here"
29
+
30
+ # Generate text
31
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
32
+ output = model.generate(**inputs, max_new_tokens=1024)
33
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
34
+ print(response)
35
+ ```
model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f10c7d5c4d6a534bd3e055aee7715169f58cddf5aafaa526e09ea0b61592aa5
3
+ size 17717352110
training_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model": "mistralai/Mistral-7B-Instruct-v0.3",
3
+ "finetuning_type": "LoRA",
4
+ "lora_r": 16,
5
+ "lora_alpha": 32,
6
+ "lora_dropout": 0.05,
7
+ "quantize": "bnb.nf4",
8
+ "context_length": 4098,
9
+ "train_batch_size": 4,
10
+ "learning_rate": "2e-4",
11
+ "train_steps": 2000
12
+ }