TokenBender commited on
Commit
badeea0
1 Parent(s): 931deff

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +50 -0
README.md ADDED
@@ -0,0 +1,50 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+
3
+ ### Overview:
4
+ description:
5
+
6
+ This is a llama2 7B HF chat model fine-tuned on 122k code instructions. In my early experiments it seems to be doing very well.
7
+
8
+ additional_info:
9
+
10
+ It's a bottom of the barrel model 😂 but after quantization it can be
11
+ valuable for sure. It definitely proves that a 7B can be useful for boilerplate
12
+ code stuff though.
13
+
14
+ ### Plans:
15
+ next_steps: "I've a few things in mind and after that this will be more valuable."
16
+
17
+ tasks:
18
+
19
+ - name: "I'll quantize these"
20
+ timeline: "Possibly tonight or tomorrow in the day"
21
+ result: "Then it can be run locally with 4G ram."
22
+ - name: "I've used alpaca style instruction tuning"
23
+ improvement: |
24
+ I'll switch to llama2 style [INST]<<SYS>> style and see if
25
+ it improves anything.
26
+ - name: "HumanEval report and checking for any training data leaks"
27
+ - attempt: "I'll try 8k context via RoPE enhancement"
28
+ hypothesis: "Let's see if that degrades performance or not."
29
+ commercial_use: |
30
+ So far I think this can be used commercially but this is a adapter on Meta's llama2 with
31
+ some gating issues so that is there.
32
+ contact_info: "If you find any issues or want to just holler at me, you can reach out to me - https://twitter.com/4evaBehindSOTA"
33
+
34
+ ### Library:
35
+ name: "peft"
36
+
37
+ ### Training procedure:
38
+ quantization_config:
39
+ load_in_8bit: False
40
+ load_in_4bit: True
41
+ llm_int8_threshold: 6.0
42
+ llm_int8_skip_modules: None
43
+ llm_int8_enable_fp32_cpu_offload: False
44
+ llm_int8_has_fp16_weight: False
45
+ bnb_4bit_quant_type: "nf4"
46
+ bnb_4bit_use_double_quant: False
47
+ bnb_4bit_compute_dtype: "float16"
48
+
49
+ ### Framework versions:
50
+ PEFT: "0.5.0.dev0"