Commit
Β·
9a0d878
1
Parent(s):
dc3e04f
prepare for exllamav3
Browse files- README.md +14 -13
- {auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/INSTRUCTIONS.txt +0 -0
- {auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/auto-exl2-upload.zip +0 -0
- {auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/exl2-quant.py +0 -0
- {auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/linux-setup.sh +0 -0
- {auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/windows-setup.bat +0 -0
- {exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/INSTRUCTIONS.txt +0 -0
- {exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/exl2-multi-quant-local.zip +0 -0
- {exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/exl2-quant.py +0 -0
- {exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/linux-setup.sh +0 -0
- {exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/windows-setup.bat +0 -0
- exllamav3 scripts/placeholder +0 -0
README.md
CHANGED
|
@@ -11,24 +11,25 @@ Feel free to send in PRs or use this code however you'd like.\
|
|
| 11 |
|
| 12 |
**For GitHub**: Would recommend creating pull requests and discussions on the [offical huggingface repo](https://huggingface.co/Anthonyg5005/hf-scripts)
|
| 13 |
|
| 14 |
-
##
|
| 15 |
|
| 16 |
-
- [Auto EXL2 HF upload](https://huggingface.co/Anthonyg5005/hf-scripts/resolve/main/auto-exl2-upload/auto-exl2-upload.zip?download=true)
|
| 17 |
|
| 18 |
-
- [EXL2 Local Quants](https://huggingface.co/Anthonyg5005/hf-scripts/resolve/main/exl2-multi-quant-local/exl2-multi-quant-local.zip?download=true)
|
| 19 |
-
|
| 20 |
-
- [Upload folder to HF](https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/upload%20folder%20to%20repo.py)
|
| 21 |
|
| 22 |
- [Manage branches (create/delete)](https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/manage%20branches.py)
|
| 23 |
|
|
|
|
|
|
|
| 24 |
- [EXL2 Single Quant V3](https://colab.research.google.com/#fileId=https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/ipynb/EXL2_Private_Quant_V3.ipynb) **(COLAB)**
|
| 25 |
|
| 26 |
-
|
| 27 |
|
| 28 |
-
|
| 29 |
-
|
| 30 |
-
- Allow using finegrained tokens to login scripts
|
| 31 |
|
|
|
|
|
|
|
| 32 |
## other recommended stuff
|
| 33 |
|
| 34 |
- [Exllama Discord server](https://discord.gg/NSFwVuCjRq)
|
|
@@ -44,14 +45,14 @@ Feel free to send in PRs or use this code however you'd like.\
|
|
| 44 |
- EXL2 Local Quants
|
| 45 |
- Easily creates environment to quantize models to exl2 to your local machine. Supports both Windows and Linux.
|
| 46 |
|
|
|
|
|
|
|
|
|
|
| 47 |
- Upload folder to repo
|
| 48 |
-
- Uploads user specified folder to specified repo, can create private repos too. Not the same as git commit and push, instead uploads any additional files. This is more of a practice for me than for actual usage.
|
| 49 |
|
| 50 |
- Manage branches
|
| 51 |
- Run script and follow prompts. You will be required to be logged in to HF Hub. If you are not logged in, you will need a WRITE token. You can get one in your [HuggingFace settings](https://huggingface.co/settings/tokens). Colab and Kaggle secret keys are supported.
|
| 52 |
-
|
| 53 |
-
- EXL2 Single Quant
|
| 54 |
-
- Allows you to quantize to exl2 using colab. This version creates a exl2 quant to upload to private repo. Only 7B tested on colab.
|
| 55 |
|
| 56 |
- Download models (oobabooga)
|
| 57 |
- To use the script, open a terminal and run '`python download-model.py USER/MODEL:BRANCH`'. There's also a '`--help`' flag to show the available arguments. To download from private repositories, make sure to login using '`huggingface-cli login`' or (not recommended) `HF_TOKEN` environment variable.
|
|
|
|
| 11 |
|
| 12 |
**For GitHub**: Would recommend creating pull requests and discussions on the [offical huggingface repo](https://huggingface.co/Anthonyg5005/hf-scripts)
|
| 13 |
|
| 14 |
+
## main files
|
| 15 |
|
| 16 |
+
- [Auto EXL2 HF upload](https://huggingface.co/Anthonyg5005/hf-scripts/resolve/main/exllamav2%20scripts/auto-exl2-upload/auto-exl2-upload.zip?download=true)
|
| 17 |
|
| 18 |
+
- [EXL2 Local Quants](https://huggingface.co/Anthonyg5005/hf-scripts/resolve/main/exllamav2%20scripts/exl2-multi-quant-local/exl2-multi-quant-local.zip?download=true)
|
|
|
|
|
|
|
| 19 |
|
| 20 |
- [Manage branches (create/delete)](https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/manage%20branches.py)
|
| 21 |
|
| 22 |
+
## outdated or not main focus
|
| 23 |
+
|
| 24 |
- [EXL2 Single Quant V3](https://colab.research.google.com/#fileId=https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/ipynb/EXL2_Private_Quant_V3.ipynb) **(COLAB)**
|
| 25 |
|
| 26 |
+
- [Upload folder to HF](https://huggingface.co/Anthonyg5005/hf-scripts/blob/main/upload%20folder%20to%20repo.py)
|
| 27 |
|
| 28 |
+
<!--
|
| 29 |
+
## work in progress/not tested (ordered by priority)
|
|
|
|
| 30 |
|
| 31 |
+
none for now. perhaps adding finegrained token support to my hf login code
|
| 32 |
+
-->
|
| 33 |
## other recommended stuff
|
| 34 |
|
| 35 |
- [Exllama Discord server](https://discord.gg/NSFwVuCjRq)
|
|
|
|
| 45 |
- EXL2 Local Quants
|
| 46 |
- Easily creates environment to quantize models to exl2 to your local machine. Supports both Windows and Linux.
|
| 47 |
|
| 48 |
+
- EXL2 Single Quant
|
| 49 |
+
- Allows you to quantize to exl2 using colab. This version creates a exl2 quant to upload to private repo. Only 7B tested on colab.
|
| 50 |
+
|
| 51 |
- Upload folder to repo
|
| 52 |
+
- Uploads user specified folder to specified repo, can create private repos too. Not the same as git commit and push, instead uploads any additional files. This is more of a practice for me than for actual usage as most of the time it crashes on the quantizing process due to lack of ram.
|
| 53 |
|
| 54 |
- Manage branches
|
| 55 |
- Run script and follow prompts. You will be required to be logged in to HF Hub. If you are not logged in, you will need a WRITE token. You can get one in your [HuggingFace settings](https://huggingface.co/settings/tokens). Colab and Kaggle secret keys are supported.
|
|
|
|
|
|
|
|
|
|
| 56 |
|
| 57 |
- Download models (oobabooga)
|
| 58 |
- To use the script, open a terminal and run '`python download-model.py USER/MODEL:BRANCH`'. There's also a '`--help`' flag to show the available arguments. To download from private repositories, make sure to login using '`huggingface-cli login`' or (not recommended) `HF_TOKEN` environment variable.
|
{auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/INSTRUCTIONS.txt
RENAMED
|
File without changes
|
{auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/auto-exl2-upload.zip
RENAMED
|
File without changes
|
{auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/exl2-quant.py
RENAMED
|
File without changes
|
{auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/linux-setup.sh
RENAMED
|
File without changes
|
{auto-exl2-upload β exllamav2 scripts/auto-exl2-upload}/windows-setup.bat
RENAMED
|
File without changes
|
{exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/INSTRUCTIONS.txt
RENAMED
|
File without changes
|
{exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/exl2-multi-quant-local.zip
RENAMED
|
File without changes
|
{exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/exl2-quant.py
RENAMED
|
File without changes
|
{exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/linux-setup.sh
RENAMED
|
File without changes
|
{exl2-multi-quant-local β exllamav2 scripts/exl2-multi-quant-local}/windows-setup.bat
RENAMED
|
File without changes
|
exllamav3 scripts/placeholder
ADDED
|
File without changes
|