Commit
·
d7e42a3
1
Parent(s):
7e92fcf
add optional dev build
Browse files
auto-exl2-upload/auto-exl2-upload.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dd9f20f25139aba9a0e0f4742f8d9852a597e8877ce8f5fa05c7abe18cc77f1f
|
| 3 |
+
size 7732
|
auto-exl2-upload/linux-setup.sh
CHANGED
|
@@ -22,6 +22,14 @@ else
|
|
| 22 |
exit
|
| 23 |
fi
|
| 24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
| 26 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
| 27 |
|
|
@@ -49,7 +57,11 @@ rm enter-venv.sh
|
|
| 49 |
|
| 50 |
# download stuff
|
| 51 |
echo "Downloading files"
|
| 52 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 55 |
|
|
|
|
| 22 |
exit
|
| 23 |
fi
|
| 24 |
|
| 25 |
+
# ask for exllamav2 version
|
| 26 |
+
read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
|
| 27 |
+
if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
|
| 28 |
+
echo "Invalid version of exllama. Please enter stable or dev."
|
| 29 |
+
read -p "Press enter to continue"
|
| 30 |
+
exit
|
| 31 |
+
fi
|
| 32 |
+
|
| 33 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
| 34 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
| 35 |
|
|
|
|
| 57 |
|
| 58 |
# download stuff
|
| 59 |
echo "Downloading files"
|
| 60 |
+
if [ "$exllamav2_version" = "stable" ]; then
|
| 61 |
+
git clone https://github.com/turboderp/exllamav2
|
| 62 |
+
elif [ "$exllamav2_version" = "dev" ]; then
|
| 63 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
| 64 |
+
fi
|
| 65 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 67 |
|
auto-exl2-upload/windows-setup.bat
CHANGED
|
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
|
|
| 24 |
exit
|
| 25 |
)
|
| 26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
| 28 |
echo CUDA compilers:
|
| 29 |
where nvcc
|
|
@@ -41,6 +49,7 @@ if "%cuda_version%"=="11" (
|
|
| 41 |
exit
|
| 42 |
)
|
| 43 |
|
|
|
|
| 44 |
echo Deleting potential conflicting files
|
| 45 |
del convert-to-safetensors.py
|
| 46 |
del download-model.py
|
|
@@ -50,7 +59,11 @@ del enter-venv.sh
|
|
| 50 |
|
| 51 |
REM download stuff
|
| 52 |
echo Downloading files...
|
| 53 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 55 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 56 |
|
|
|
|
| 24 |
exit
|
| 25 |
)
|
| 26 |
|
| 27 |
+
REM ask for exllamav2 version
|
| 28 |
+
set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
|
| 29 |
+
if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
|
| 30 |
+
echo Invalid exllamav2 version. Please enter stable or dev.
|
| 31 |
+
pause
|
| 32 |
+
exit
|
| 33 |
+
)
|
| 34 |
+
|
| 35 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
| 36 |
echo CUDA compilers:
|
| 37 |
where nvcc
|
|
|
|
| 49 |
exit
|
| 50 |
)
|
| 51 |
|
| 52 |
+
|
| 53 |
echo Deleting potential conflicting files
|
| 54 |
del convert-to-safetensors.py
|
| 55 |
del download-model.py
|
|
|
|
| 59 |
|
| 60 |
REM download stuff
|
| 61 |
echo Downloading files...
|
| 62 |
+
if "%exllamav2_version%"=="stable" (
|
| 63 |
+
git clone https://github.com/turboderp/exllamav2
|
| 64 |
+
) else if "%exllamav2_version%"=="dev" (
|
| 65 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
| 66 |
+
)
|
| 67 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 68 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 69 |
|
exl2-multi-quant-local/exl2-multi-quant-local.zip
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:95a2bf855195c1b9c45d79b550c36dae3dc0f833a51b782694053c1fcccd14ec
|
| 3 |
+
size 6768
|
exl2-multi-quant-local/linux-setup.sh
CHANGED
|
@@ -22,6 +22,14 @@ else
|
|
| 22 |
exit
|
| 23 |
fi
|
| 24 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 25 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
| 26 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
| 27 |
|
|
@@ -49,7 +57,11 @@ rm enter-venv.sh
|
|
| 49 |
|
| 50 |
# download stuff
|
| 51 |
echo "Downloading files"
|
| 52 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 55 |
|
|
|
|
| 22 |
exit
|
| 23 |
fi
|
| 24 |
|
| 25 |
+
# ask for exllamav2 version
|
| 26 |
+
read -p "Want to build stable or dev version of exllamav2? (stable, dev): " exllamav2_version
|
| 27 |
+
if [ "$exllamav2_version" != "stable" ] && [ "$exllamav2_version" != "dev" ]; then
|
| 28 |
+
echo "Invalid version of exllama. Please enter stable or dev."
|
| 29 |
+
read -p "Press enter to continue"
|
| 30 |
+
exit
|
| 31 |
+
fi
|
| 32 |
+
|
| 33 |
# if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8. If ROCm, install pytorch for ROCm 5.7
|
| 34 |
read -p "Please enter your GPU compute version, CUDA 11/12 or AMD ROCm (11, 12, rocm): " pytorch_version
|
| 35 |
|
|
|
|
| 57 |
|
| 58 |
# download stuff
|
| 59 |
echo "Downloading files"
|
| 60 |
+
if [ "$exllamav2_version" = "stable" ]; then
|
| 61 |
+
git clone https://github.com/turboderp/exllamav2
|
| 62 |
+
elif [ "$exllamav2_version" = "dev" ]; then
|
| 63 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
| 64 |
+
fi
|
| 65 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 67 |
|
exl2-multi-quant-local/windows-setup.bat
CHANGED
|
@@ -24,6 +24,14 @@ if "%gitwget%"=="y" (
|
|
| 24 |
exit
|
| 25 |
)
|
| 26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
| 28 |
echo CUDA compilers:
|
| 29 |
where nvcc
|
|
@@ -50,7 +58,11 @@ del enter-venv.sh
|
|
| 50 |
|
| 51 |
REM download stuff
|
| 52 |
echo Downloading files...
|
| 53 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 55 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 56 |
|
|
|
|
| 24 |
exit
|
| 25 |
)
|
| 26 |
|
| 27 |
+
REM ask for exllamav2 version
|
| 28 |
+
set /p exllamav2_version="Would you like to build stable or dev version of exllamav2? (stable, dev): "
|
| 29 |
+
if not "%exllamav2_version%"=="stable" if not "%exllamav2_version%"=="dev" (
|
| 30 |
+
echo Invalid exllamav2 version. Please enter stable or dev.
|
| 31 |
+
pause
|
| 32 |
+
exit
|
| 33 |
+
)
|
| 34 |
+
|
| 35 |
REM if CUDA version 12 install pytorch for 12.1, else if CUDA 11 install pytorch for 11.8
|
| 36 |
echo CUDA compilers:
|
| 37 |
where nvcc
|
|
|
|
| 58 |
|
| 59 |
REM download stuff
|
| 60 |
echo Downloading files...
|
| 61 |
+
if "%exllamav2_version%"=="stable" (
|
| 62 |
+
git clone https://github.com/turboderp/exllamav2
|
| 63 |
+
) else if "%exllamav2_version%"=="dev" (
|
| 64 |
+
git clone https://github.com/turboderp/exllamav2 -b dev
|
| 65 |
+
)
|
| 66 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/convert-to-safetensors.py
|
| 67 |
wget https://raw.githubusercontent.com/oobabooga/text-generation-webui/main/download-model.py
|
| 68 |
|