Anthonyg5005
commited on
Commit
•
d7e42a3
1
Parent(s):
7e92fcf
add optional dev build
Browse files
auto-exl2-upload/auto-exl2-upload.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd9f20f25139aba9a0e0f4742f8d9852a597e8877ce8f5fa05c7abe18cc77f1f
|
3 |
+
size 7732
|
auto-exl2-upload/linux-setup.sh
CHANGED
@@ -22,6 +22,14 @@ else
|
|
22 |
exit
|
23 |
fi
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
26 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
27 |
|
@@ -49,7 +57,11 @@ rm enter-venv.sh
|
|
49 |
|
50 |
# download stuff
|
51 |
echo "Downloading files"
|
52 |
-
|
|
|
|
|
|
|
|
|
53 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
55 |
|
|
|
22 |
exit
|
23 |
fi
|
24 |
|
25 |
+
# ask for exllamav2 version
|
26 |
+
read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
|
27 |
+
if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
|
28 |
+
echo "Invalid version of exllama. Please enter stable or dev."
|
29 |
+
read -p "Press enter to continue"
|
30 |
+
exit
|
31 |
+
fi
|
32 |
+
|
33 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
34 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
35 |
|
|
|
57 |
|
58 |
# download stuff
|
59 |
echo "Downloading files"
|
60 |
+
if [ "$exllamav2_version" = "stable" ]; then
|
61 |
+
git clone https://github.com/turboderp/exllamav2
|
62 |
+
elif [ "$exllamav2_version" = "dev" ]; then
|
63 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
64 |
+
fi
|
65 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
67 |
|
auto-exl2-upload/windows-setup.bat
CHANGED
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
|
|
24 |
exit
|
25 |
)
|
26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
28 |
echo CUDA compilers:
|
29 |
where nvcc
|
@@ -41,6 +49,7 @@ if "%cuda_version%"=="11" (
|
|
41 |
exit
|
42 |
)
|
43 |
|
|
|
44 |
echo Deleting potential conflicting files
|
45 |
del convert-to-safetensors.py
|
46 |
del download-model.py
|
@@ -50,7 +59,11 @@ del enter-venv.sh
|
|
50 |
|
51 |
REM download stuff
|
52 |
echo Downloading files...
|
53 |
-
|
|
|
|
|
|
|
|
|
54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
55 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
56 |
|
|
|
24 |
exit
|
25 |
)
|
26 |
|
27 |
+
REM ask for exllamav2 version
|
28 |
+
set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
|
29 |
+
if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
|
30 |
+
echo Invalid exllamav2 version. Please enter stable or dev.
|
31 |
+
pause
|
32 |
+
exit
|
33 |
+
)
|
34 |
+
|
35 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
36 |
echo CUDA compilers:
|
37 |
where nvcc
|
|
|
49 |
exit
|
50 |
)
|
51 |
|
52 |
+
|
53 |
echo Deleting potential conflicting files
|
54 |
del convert-to-safetensors.py
|
55 |
del download-model.py
|
|
|
59 |
|
60 |
REM download stuff
|
61 |
echo Downloading files...
|
62 |
+
if "%exllamav2_version%"=="stable" (
|
63 |
+
git clone https://github.com/turboderp/exllamav2
|
64 |
+
) else if "%exllamav2_version%"=="dev" (
|
65 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
66 |
+
)
|
67 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
68 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
69 |
|
exl2-multi-quant-local/exl2-multi-quant-local.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95a2bf855195c1b9c45d79b550c36dae3dc0f833a51b782694053c1fcccd14ec
|
3 |
+
size 6768
|
exl2-multi-quant-local/linux-setup.sh
CHANGED
@@ -22,6 +22,14 @@ else
|
|
22 |
exit
|
23 |
fi
|
24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
26 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
27 |
|
@@ -49,7 +57,11 @@ rm enter-venv.sh
|
|
49 |
|
50 |
# download stuff
|
51 |
echo "Downloading files"
|
52 |
-
|
|
|
|
|
|
|
|
|
53 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
55 |
|
|
|
22 |
exit
|
23 |
fi
|
24 |
|
25 |
+
# ask for exllamav2 version
|
26 |
+
read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
|
27 |
+
if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
|
28 |
+
echo "Invalid version of exllama. Please enter stable or dev."
|
29 |
+
read -p "Press enter to continue"
|
30 |
+
exit
|
31 |
+
fi
|
32 |
+
|
33 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
34 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
35 |
|
|
|
57 |
|
58 |
# download stuff
|
59 |
echo "Downloading files"
|
60 |
+
if [ "$exllamav2_version" = "stable" ]; then
|
61 |
+
git clone https://github.com/turboderp/exllamav2
|
62 |
+
elif [ "$exllamav2_version" = "dev" ]; then
|
63 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
64 |
+
fi
|
65 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
67 |
|
exl2-multi-quant-local/windows-setup.bat
CHANGED
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
|
|
24 |
exit
|
25 |
)
|
26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
28 |
echo CUDA compilers:
|
29 |
where nvcc
|
@@ -50,7 +58,11 @@ del enter-venv.sh
|
|
50 |
|
51 |
REM download stuff
|
52 |
echo Downloading files...
|
53 |
-
|
|
|
|
|
|
|
|
|
54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
55 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
56 |
|
|
|
24 |
exit
|
25 |
)
|
26 |
|
27 |
+
REM ask for exllamav2 version
|
28 |
+
set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
|
29 |
+
if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
|
30 |
+
echo Invalid exllamav2 version. Please enter stable or dev.
|
31 |
+
pause
|
32 |
+
exit
|
33 |
+
)
|
34 |
+
|
35 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
36 |
echo CUDA compilers:
|
37 |
where nvcc
|
|
|
58 |
|
59 |
REM download stuff
|
60 |
echo Downloading files...
|
61 |
+
if "%exllamav2_version%"=="stable" (
|
62 |
+
git clone https://github.com/turboderp/exllamav2
|
63 |
+
) else if "%exllamav2_version%"=="dev" (
|
64 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
65 |
+
)
|
66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
67 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
68 |
|