Anthonyg5005 commited on
Commit
d7e42a3
1 Parent(s): 7e92fcf

add optional dev build

Browse files
auto-exl2-upload/auto-exl2-upload.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a253a898953cc1283cd0950308174d669c5b568a0abcdb2f322ccf49203ad22b
3
- size 7528
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd9f20f25139aba9a0e0f4742f8d9852a597e8877ce8f5fa05c7abe18cc77f1f
3
+ size 7732
auto-exl2-upload/linux-setup.sh CHANGED
@@ -22,6 +22,14 @@ else
22
  exit
23
  fi
24
 
 
 
 
 
 
 
 
 
25
  # if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
26
  read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
27
 
@@ -49,7 +57,11 @@ rm enter-venv.sh
49
 
50
  # download stuff
51
  echo "Downloading files"
52
- git clone https://github.com/turboderp/exllamav2
 
 
 
 
53
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
54
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
55
 
 
22
  exit
23
  fi
24
 
25
+ # ask for exllamav2 version
26
+ read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
27
+ if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
28
+ echo "Invalid version of exllama. Please enter stable or dev."
29
+ read -p "Press enter to continue"
30
+ exit
31
+ fi
32
+
33
  # if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
34
  read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
35
 
 
57
 
58
  # download stuff
59
  echo "Downloading files"
60
+ if [ "$exllamav2_version" = "stable" ]; then
61
+ git clone https://github.com/turboderp/exllamav2
62
+ elif [ "$exllamav2_version" = "dev" ]; then
63
+ git clone https://github.com/turboderp/exllamav2 -b dev
64
+ fi
65
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
66
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
67
 
auto-exl2-upload/windows-setup.bat CHANGED
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
24
  exit
25
  )
26
 
 
 
 
 
 
 
 
 
27
  REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
28
  echo CUDA compilers:
29
  where nvcc
@@ -41,6 +49,7 @@ if "%cuda_version%"=="11" (
41
  exit
42
  )
43
 
 
44
  echo Deleting potential conflicting files
45
  del convert-to-safetensors.py
46
  del download-model.py
@@ -50,7 +59,11 @@ del enter-venv.sh
50
 
51
  REM download stuff
52
  echo Downloading files...
53
- git clone https://github.com/turboderp/exllamav2
 
 
 
 
54
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
55
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
56
 
 
24
  exit
25
  )
26
 
27
+ REM ask for exllamav2 version
28
+ set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
29
+ if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
30
+ echo Invalid exllamav2 version. Please enter stable or dev.
31
+ pause
32
+ exit
33
+ )
34
+
35
  REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
36
  echo CUDA compilers:
37
  where nvcc
 
49
  exit
50
  )
51
 
52
+
53
  echo Deleting potential conflicting files
54
  del convert-to-safetensors.py
55
  del download-model.py
 
59
 
60
  REM download stuff
61
  echo Downloading files...
62
+ if "%exllamav2_version%"=="stable" (
63
+ git clone https://github.com/turboderp/exllamav2
64
+ ) else if "%exllamav2_version%"=="dev" (
65
+ git clone https://github.com/turboderp/exllamav2 -b dev
66
+ )
67
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
68
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
69
 
exl2-multi-quant-local/exl2-multi-quant-local.zip CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a99d13ce17a3d436d50d80dc4a9e9bf784ef77efb400190233c715c86f936933
3
- size 6563
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95a2bf855195c1b9c45d79b550c36dae3dc0f833a51b782694053c1fcccd14ec
3
+ size 6768
exl2-multi-quant-local/linux-setup.sh CHANGED
@@ -22,6 +22,14 @@ else
22
  exit
23
  fi
24
 
 
 
 
 
 
 
 
 
25
  # if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
26
  read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
27
 
@@ -49,7 +57,11 @@ rm enter-venv.sh
49
 
50
  # download stuff
51
  echo "Downloading files"
52
- git clone https://github.com/turboderp/exllamav2
 
 
 
 
53
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
54
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
55
 
 
22
  exit
23
  fi
24
 
25
+ # ask for exllamav2 version
26
+ read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
27
+ if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
28
+ echo "Invalid version of exllama. Please enter stable or dev."
29
+ read -p "Press enter to continue"
30
+ exit
31
+ fi
32
+
33
  # if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
34
  read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
35
 
 
57
 
58
  # download stuff
59
  echo "Downloading files"
60
+ if [ "$exllamav2_version" = "stable" ]; then
61
+ git clone https://github.com/turboderp/exllamav2
62
+ elif [ "$exllamav2_version" = "dev" ]; then
63
+ git clone https://github.com/turboderp/exllamav2 -b dev
64
+ fi
65
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
66
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
67
 
exl2-multi-quant-local/windows-setup.bat CHANGED
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
24
  exit
25
  )
26
 
 
 
 
 
 
 
 
 
27
  REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
28
  echo CUDA compilers:
29
  where nvcc
@@ -50,7 +58,11 @@ del enter-venv.sh
50
 
51
  REM download stuff
52
  echo Downloading files...
53
- git clone https://github.com/turboderp/exllamav2
 
 
 
 
54
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
55
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
56
 
 
24
  exit
25
  )
26
 
27
+ REM ask for exllamav2 version
28
+ set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
29
+ if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
30
+ echo Invalid exllamav2 version. Please enter stable or dev.
31
+ pause
32
+ exit
33
+ )
34
+
35
  REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
36
  echo CUDA compilers:
37
  where nvcc
 
58
 
59
  REM download stuff
60
  echo Downloading files...
61
+ if "%exllamav2_version%"=="stable" (
62
+ git clone https://github.com/turboderp/exllamav2
63
+ ) else if "%exllamav2_version%"=="dev" (
64
+ git clone https://github.com/turboderp/exllamav2 -b dev
65
+ )
66
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
67
  wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
68