Spaces:
Sleeping
Sleeping
Switch model from llama to a small 1.5B Qwen
Browse files
app.py
CHANGED
|
@@ -4,7 +4,7 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
| 4 |
import plotly.express as px
|
| 5 |
|
| 6 |
|
| 7 |
-
model_name = '
|
| 8 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 9 |
|
| 10 |
@st.cache_resource
|
|
|
|
| 4 |
import plotly.express as px
|
| 5 |
|
| 6 |
|
| 7 |
+
model_name = 'Qwen/Qwen2-1.5B'
|
| 8 |
device = 'cuda' if torch.cuda.is_available() else 'cpu'
|
| 9 |
|
| 10 |
@st.cache_resource
|