Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -56,7 +56,7 @@ if not torch.cuda.is_available():
|
|
56 |
if torch.cuda.is_available():
|
57 |
|
58 |
# Model and Tokenizer Configuration
|
59 |
-
model_id = "meta-llama/Llama-3.1-8B
|
60 |
bnb_config = BitsAndBytesConfig(
|
61 |
load_in_4bit=True,
|
62 |
bnb_4bit_use_double_quant=False,
|
@@ -64,7 +64,7 @@ if torch.cuda.is_available():
|
|
64 |
bnb_4bit_compute_dtype=torch.bfloat16
|
65 |
)
|
66 |
base_model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", quantization_config=bnb_config)
|
67 |
-
model = PeftModel.from_pretrained(base_model, "ranamhamoud/
|
68 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
69 |
tokenizer.pad_token = tokenizer.eos_token
|
70 |
|
|
|
56 |
if torch.cuda.is_available():
|
57 |
|
58 |
# Model and Tokenizer Configuration
|
59 |
+
model_id = "meta-llama/Llama-3.1-8B"
|
60 |
bnb_config = BitsAndBytesConfig(
|
61 |
load_in_4bit=True,
|
62 |
bnb_4bit_use_double_quant=False,
|
|
|
64 |
bnb_4bit_compute_dtype=torch.bfloat16
|
65 |
)
|
66 |
base_model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", quantization_config=bnb_config)
|
67 |
+
model = PeftModel.from_pretrained(base_model, "ranamhamoud/storytellai-2.0")
|
68 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
69 |
tokenizer.pad_token = tokenizer.eos_token
|
70 |
|