Commit
·
ec9b8d8
1
Parent(s):
9a0d878
6hb default again
Browse filesmuch bigger vram usage in newer models, not really worth it much anymore.
also removed fast safetensors because deprecated
exllamav2 scripts/auto-exl2-upload/auto-exl2-upload.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d5c346d6dd344f29c196b33bdb8e2d07cf8d70e1e9135e1f2bfab8550bcba4e
|
3 |
+
size 8717
|
exllamav2 scripts/auto-exl2-upload/exl2-quant.py
CHANGED
@@ -203,7 +203,7 @@ for bpw in bpwvalue:
|
|
203 |
os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
204 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
205 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
206 |
-
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip}
|
207 |
if result.returncode != 0:
|
208 |
print("Quantization failed.")
|
209 |
sys.exit("Exiting...")
|
|
|
203 |
os.makedirs(f"{model}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
204 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
205 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
206 |
+
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {model}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
207 |
if result.returncode != 0:
|
208 |
print("Quantization failed.")
|
209 |
sys.exit("Exiting...")
|
exllamav2 scripts/exl2-multi-quant-local/exl2-multi-quant-local.zip
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd45a8c909552d9e4f1f8b2069d210e91a2ce0fd8043b45dcffef764dab370ba
|
3 |
+
size 7513
|
exllamav2 scripts/exl2-multi-quant-local/exl2-quant.py
CHANGED
@@ -150,7 +150,7 @@ for bpw in bpwvalue:
|
|
150 |
os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
151 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
152 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
153 |
-
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip}
|
154 |
if result.returncode != 0:
|
155 |
print("Quantization failed.")
|
156 |
sys.exit("Exiting...")
|
|
|
150 |
os.makedirs(f"{modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw", exist_ok=True) #create compile full directory
|
151 |
subprocess.run(f"{oscp} models{slsh}{model}{slsh}config.json {model}-exl2-{bpw}bpw-WD", shell=True) #copy config to working directory
|
152 |
#more settings exist in the convert.py script, to veiw them go to docs/convert.md or https://github.com/turboderp/exllamav2/blob/master/doc/convert.md
|
153 |
+
result = subprocess.run(f"{pyt} exllamav2/convert.py -i models/{model} -o {model}-exl2-{bpw}bpw-WD -cf {modelname}-exl2-quants{slsh}{modelname}-exl2-{bpw}bpw -b {bpw}{mskip}", shell=True) #run quantization and exit if failed (Credit to turbo for his dedication to exl2)
|
154 |
if result.returncode != 0:
|
155 |
print("Quantization failed.")
|
156 |
sys.exit("Exiting...")
|