Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -6,7 +6,6 @@ import logging
|
|
6 |
import sys
|
7 |
import os
|
8 |
from accelerate import infer_auto_device_map, init_empty_weights
|
9 |
-
import bitsandbytes as bnb
|
10 |
|
11 |
# Configure logging
|
12 |
logging.basicConfig(
|
@@ -46,20 +45,14 @@ try:
|
|
46 |
tokenizer.pad_token = tokenizer.eos_token
|
47 |
logger.info("Tokenizer loaded successfully")
|
48 |
|
49 |
-
# Load model with
|
50 |
logger.info("Loading model...")
|
51 |
model = AutoModelForCausalLM.from_pretrained(
|
52 |
model_name,
|
53 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32,
|
54 |
trust_remote_code=True,
|
55 |
token=hf_token,
|
56 |
-
device_map="auto"
|
57 |
-
quantization_config={
|
58 |
-
"load_in_4bit": True,
|
59 |
-
"bnb_4bit_compute_dtype": torch.float16,
|
60 |
-
"bnb_4bit_use_double_quant": True,
|
61 |
-
"bnb_4bit_quant_type": "nf4"
|
62 |
-
}
|
63 |
)
|
64 |
logger.info("Model loaded successfully")
|
65 |
|
|
|
6 |
import sys
|
7 |
import os
|
8 |
from accelerate import infer_auto_device_map, init_empty_weights
|
|
|
9 |
|
10 |
# Configure logging
|
11 |
logging.basicConfig(
|
|
|
45 |
tokenizer.pad_token = tokenizer.eos_token
|
46 |
logger.info("Tokenizer loaded successfully")
|
47 |
|
48 |
+
# Load model with basic configuration
|
49 |
logger.info("Loading model...")
|
50 |
model = AutoModelForCausalLM.from_pretrained(
|
51 |
model_name,
|
52 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32,
|
53 |
trust_remote_code=True,
|
54 |
token=hf_token,
|
55 |
+
device_map="auto"
|
|
|
|
|
|
|
|
|
|
|
|
|
56 |
)
|
57 |
logger.info("Model loaded successfully")
|
58 |
|