Spaces:
Sleeping
Sleeping
Switch model from llama to a small 1.5B Qwen
Browse files
app.py
CHANGED
@@ -4,7 +4,7 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
4 |
import plotly.express as px
|
5 |
|
6 |
|
7 |
-
model_name = '
|
8 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
9 |
|
10 |
@st.cache_resource
|
|
|
4 |
import plotly.express as px
|
5 |
|
6 |
|
7 |
+
model_name = 'Qwen/Qwen2-1.5B'
|
8 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
9 |
|
10 |
@st.cache_resource
|