File size: 293 Bytes
417b435
 
 
 
 
 
 
1
2
3
4
5
6
7
Made by merging the following lora:
https://huggingface.co/Neko-Institute-of-Science/VicUnLocked-30b-LoRA

Then quantizing with ooba's old CUDA branch of GPTQ
```
python llama.py vicunlocked-30b c4 --wbits 4 --true-sequential --act-order --save_safetensors vicunlocked-30b-4bit.safetensors
```