wilmerhenao commited on
Commit
fb7211b
1 Parent(s): 2bb239d

Introducing Olinguito: A Language Model Fine-tuned with LORA Algorithm on Alpaca-cleaned Data

Browse files

This commit adds Olinguito, a new language model derived from Dolly, which has been fine-tuned using the LORA (Low-Rank Adaptation of Large Language Models ) algorithm. Olinguito's training data has undergone meticulous cleaning, specifically extracted from Alpaca sources. By applying LORA, Olinguito aims to provide enhanced performance, accuracy, and robustness in natural language processing tasks. This commit lays the foundation for incorporating Olinguito into our Hugging Face repository, enabling users to access and utilize this refined language model for various applications.

Files changed (3) hide show
  1. README.md +42 -1
  2. adapter_config.json +20 -0
  3. adapter_model.bin +3 -0
README.md CHANGED
@@ -1,3 +1,44 @@
1
  ---
2
- license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ library_name: peft
3
  ---
4
+ ## Training procedure
5
+
6
+
7
+ The following `bitsandbytes` quantization config was used during training:
8
+ - load_in_8bit: True
9
+ - load_in_4bit: False
10
+ - llm_int8_threshold: 6.0
11
+ - llm_int8_skip_modules: None
12
+ - llm_int8_enable_fp32_cpu_offload: False
13
+ - llm_int8_has_fp16_weight: False
14
+ - bnb_4bit_quant_type: fp4
15
+ - bnb_4bit_use_double_quant: False
16
+ - bnb_4bit_compute_dtype: float32
17
+
18
+ The following `bitsandbytes` quantization config was used during training:
19
+ - load_in_8bit: True
20
+ - load_in_4bit: False
21
+ - llm_int8_threshold: 6.0
22
+ - llm_int8_skip_modules: None
23
+ - llm_int8_enable_fp32_cpu_offload: False
24
+ - llm_int8_has_fp16_weight: False
25
+ - bnb_4bit_quant_type: fp4
26
+ - bnb_4bit_use_double_quant: False
27
+ - bnb_4bit_compute_dtype: float32
28
+
29
+ The following `bitsandbytes` quantization config was used during training:
30
+ - load_in_8bit: True
31
+ - load_in_4bit: False
32
+ - llm_int8_threshold: 6.0
33
+ - llm_int8_skip_modules: None
34
+ - llm_int8_enable_fp32_cpu_offload: False
35
+ - llm_int8_has_fp16_weight: False
36
+ - bnb_4bit_quant_type: fp4
37
+ - bnb_4bit_use_double_quant: False
38
+ - bnb_4bit_compute_dtype: float32
39
+ ### Framework versions
40
+
41
+ - PEFT 0.4.0.dev0
42
+ - PEFT 0.4.0.dev0
43
+
44
+ - PEFT 0.4.0.dev0
adapter_config.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "EleutherAI/gpt-j-6B",
3
+ "bias": "none",
4
+ "fan_in_fan_out": false,
5
+ "inference_mode": true,
6
+ "init_lora_weights": true,
7
+ "layers_pattern": null,
8
+ "layers_to_transform": null,
9
+ "lora_alpha": 16,
10
+ "lora_dropout": 0.05,
11
+ "modules_to_save": null,
12
+ "peft_type": "LORA",
13
+ "r": 4,
14
+ "revision": null,
15
+ "target_modules": [
16
+ "q_proj",
17
+ "v_proj"
18
+ ],
19
+ "task_type": "CAUSAL_LM"
20
+ }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f486b4780c854cb07b402669a80fff69bf52d382be990154f4019ae53a13a25
3
+ size 7379597