Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -36,17 +36,18 @@ try:
|
|
36 |
)
|
37 |
logger.info("Tokenizer loaded successfully")
|
38 |
|
39 |
-
# Load model
|
40 |
logger.info("Loading model...")
|
41 |
model = AutoModelForCausalLM.from_pretrained(
|
42 |
model_name,
|
43 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32,
|
44 |
-
device_map="auto",
|
45 |
trust_remote_code=True
|
46 |
)
|
|
|
|
|
47 |
logger.info("Model loaded successfully")
|
48 |
|
49 |
-
# Create pipeline
|
50 |
logger.info("Creating generation pipeline...")
|
51 |
model_gen = pipeline(
|
52 |
"text-generation",
|
@@ -57,7 +58,7 @@ try:
|
|
57 |
temperature=0.7,
|
58 |
top_p=0.9,
|
59 |
repetition_penalty=1.1,
|
60 |
-
|
61 |
)
|
62 |
logger.info("Pipeline created successfully")
|
63 |
|
|
|
36 |
)
|
37 |
logger.info("Tokenizer loaded successfully")
|
38 |
|
39 |
+
# Load model - Modificada la carga del modelo
|
40 |
logger.info("Loading model...")
|
41 |
model = AutoModelForCausalLM.from_pretrained(
|
42 |
model_name,
|
43 |
torch_dtype=torch.float16 if device == "cuda" else torch.float32,
|
|
|
44 |
trust_remote_code=True
|
45 |
)
|
46 |
+
if device == "cuda":
|
47 |
+
model = model.to(device)
|
48 |
logger.info("Model loaded successfully")
|
49 |
|
50 |
+
# Create pipeline - Modificada la creaci贸n del pipeline
|
51 |
logger.info("Creating generation pipeline...")
|
52 |
model_gen = pipeline(
|
53 |
"text-generation",
|
|
|
58 |
temperature=0.7,
|
59 |
top_p=0.9,
|
60 |
repetition_penalty=1.1,
|
61 |
+
device=0 if device == "cuda" else -1
|
62 |
)
|
63 |
logger.info("Pipeline created successfully")
|
64 |
|