Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -193,7 +193,7 @@ parser.add_argument(
|
|
193 |
parser.add_argument("--base-model-path", type=str, default="lmsys/vicuna-7b-v1.3",
|
194 |
help="path of basemodel, huggingface project or local path")
|
195 |
parser.add_argument(
|
196 |
-
"--load-in-8bit", action="
|
197 |
)
|
198 |
parser.add_argument(
|
199 |
"--load-in-4bit", action="store_true", help="Use 4-bit quantization"
|
@@ -213,7 +213,7 @@ model = EaModel.from_pretrained(
|
|
213 |
torch_dtype=torch.float16,
|
214 |
low_cpu_mem_usage=True,
|
215 |
load_in_4bit=args.load_in_4bit,
|
216 |
-
load_in_8bit=
|
217 |
device_map="auto"
|
218 |
)
|
219 |
model.eval()
|
|
|
193 |
parser.add_argument("--base-model-path", type=str, default="lmsys/vicuna-7b-v1.3",
|
194 |
help="path of basemodel, huggingface project or local path")
|
195 |
parser.add_argument(
|
196 |
+
"--load-in-8bit", action="store_true", help="Use 8-bit quantization"
|
197 |
)
|
198 |
parser.add_argument(
|
199 |
"--load-in-4bit", action="store_true", help="Use 4-bit quantization"
|
|
|
213 |
torch_dtype=torch.float16,
|
214 |
low_cpu_mem_usage=True,
|
215 |
load_in_4bit=args.load_in_4bit,
|
216 |
+
load_in_8bit=True,
|
217 |
device_map="auto"
|
218 |
)
|
219 |
model.eval()
|