Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,13 +3,17 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
3 |
import pandas as pd
|
4 |
import re
|
5 |
|
6 |
-
# Load the Falcon-
|
7 |
@st.cache_resource
|
8 |
def load_falcon_model():
|
9 |
-
"""Load the Falcon-
|
10 |
-
model_name = "tiiuae/falcon-
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
12 |
-
model = AutoModelForCausalLM.from_pretrained(
|
|
|
|
|
|
|
|
|
13 |
return pipeline("text-generation", model=model, tokenizer=tokenizer)
|
14 |
|
15 |
qa_pipeline = load_falcon_model()
|
|
|
3 |
import pandas as pd
|
4 |
import re
|
5 |
|
6 |
+
# Load the Falcon-3B-Instruct model and tokenizer
|
7 |
@st.cache_resource
|
8 |
def load_falcon_model():
|
9 |
+
"""Load the Falcon-3B-Instruct model and tokenizer."""
|
10 |
+
model_name = "tiiuae/falcon-3b-instruct" # Smaller model for faster loading
|
11 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
12 |
+
model = AutoModelForCausalLM.from_pretrained(
|
13 |
+
model_name,
|
14 |
+
torch_dtype="auto", # Use FP16 if supported
|
15 |
+
device_map="auto" # Automatically distributes across available devices
|
16 |
+
)
|
17 |
return pipeline("text-generation", model=model, tokenizer=tokenizer)
|
18 |
|
19 |
qa_pipeline = load_falcon_model()
|