Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -18,12 +18,14 @@ rope_scaling = {
|
|
18 |
"factor": 8.0 # Factor de escalado
|
19 |
}
|
20 |
|
|
|
|
|
|
|
|
|
21 |
# Configurar el modelo con rope_scaling
|
22 |
config = AutoConfig.from_pretrained(model_name)
|
23 |
config.rope_scaling = rope_scaling # Aqu铆 aplicas la configuraci贸n
|
24 |
|
25 |
-
# Cargar manualmente el modelo Llama 3.1 con la configuraci贸n ajustada
|
26 |
-
model_name = "meta-llama/llama-3.1-8b-instruct"
|
27 |
model = AutoModelForCausalLM.from_pretrained(model_name, config=config)
|
28 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
29 |
|
|
|
18 |
"factor": 8.0 # Factor de escalado
|
19 |
}
|
20 |
|
21 |
+
|
22 |
+
# Cargar manualmente el modelo Llama 3.1 con la configuraci贸n ajustada
|
23 |
+
model_name = "meta-llama/llama-3.1-8b-instruct"
|
24 |
+
|
25 |
# Configurar el modelo con rope_scaling
|
26 |
config = AutoConfig.from_pretrained(model_name)
|
27 |
config.rope_scaling = rope_scaling # Aqu铆 aplicas la configuraci贸n
|
28 |
|
|
|
|
|
29 |
model = AutoModelForCausalLM.from_pretrained(model_name, config=config)
|
30 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
31 |
|