Anthonyg5005 commited on
Commit
157ee7d
1 Parent(s): 14324a5

head bits 8 default

Browse files

doesn't make much of a vram difference for better results

auto-exl2-upload/auto-exl2-upload.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:180db595eefd1f31ec9b985db95c5f3326a79ad91329ba02c7b4d6676015f4d4
3
- size 7219
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46c9216642f00007b56581890c3ca89426a6ce7148bbca4f19a5ec0a04ac5e16
3
+ size 7202
auto-exl2-upload/exl2-quant.py CHANGED
@@ -168,7 +168,7 @@ for bpw in bpwvalue:
168
  os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
169
  subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
170
  #more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
171
- result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
172
  if result.returncode != 0:
173
  print("Quantization failed.")
174
  sys.exit("Exiting...")
 
168
  os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
169
  subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
170
  #more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
171
+ result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip} -hb 8", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
172
  if result.returncode != 0:
173
  print("Quantization failed.")
174
  sys.exit("Exiting...")
exl2-multi-quant-local/exl2-multi-quant-local.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bae7404db9d1534b0eec7a865cf53dc17b29111e7107e152bde428ceb819e834
3
- size 6229
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:030dc763b103c4d58919a31ed238ff839bd0b1b0ec5dd00a0bcdd1571eed778b
3
+ size 6211
exl2-multi-quant-local/exl2-quant.py CHANGED
@@ -122,7 +122,7 @@ for bpw in bpwvalue:
122
  os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
123
  subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
124
  #more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
125
- result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
126
  if result.returncode != 0:
127
  print("Quantization failed.")
128
  sys.exit("Exiting...")
 
122
  os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
123
  subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
124
  #more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
125
+ result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip} -hb 8", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
126
  if result.returncode != 0:
127
  print("Quantization failed.")
128
  sys.exit("Exiting...")