Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -18,6 +18,12 @@ hf_hub_download(
|
|
| 18 |
local_dir = "./models"
|
| 19 |
)
|
| 20 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 21 |
hf_hub_download(
|
| 22 |
repo_id="bartowski/Reflection-Llama-3.1-70B-GGUF",
|
| 23 |
filename="Reflection-Llama-3.1-70B.imatrix",
|
|
@@ -114,7 +120,8 @@ demo = gr.ChatInterface(
|
|
| 114 |
respond,
|
| 115 |
additional_inputs=[
|
| 116 |
gr.Dropdown([
|
| 117 |
-
"Reflection-Llama-3.1-70B.Q3_K_L.gguf"
|
|
|
|
| 118 |
],
|
| 119 |
value="Reflection-Llama-3.1-70B.Q3_K_L.gguf",
|
| 120 |
label="Model"
|
|
|
|
| 18 |
local_dir = "./models"
|
| 19 |
)
|
| 20 |
|
| 21 |
+
hf_hub_download(
|
| 22 |
+
repo_id="jhofseth/Reflection-Llama-3.1-70B-GGUF",
|
| 23 |
+
filename="Reflection-Llama-3.1-70B-IQ3_XXS.gguf",
|
| 24 |
+
local_dir = "./models"
|
| 25 |
+
)
|
| 26 |
+
|
| 27 |
hf_hub_download(
|
| 28 |
repo_id="bartowski/Reflection-Llama-3.1-70B-GGUF",
|
| 29 |
filename="Reflection-Llama-3.1-70B.imatrix",
|
|
|
|
| 120 |
respond,
|
| 121 |
additional_inputs=[
|
| 122 |
gr.Dropdown([
|
| 123 |
+
"Reflection-Llama-3.1-70B.Q3_K_L.gguf",
|
| 124 |
+
"Reflection-Llama-3.1-70B-IQ3_XXS.gguf"
|
| 125 |
],
|
| 126 |
value="Reflection-Llama-3.1-70B.Q3_K_L.gguf",
|
| 127 |
label="Model"
|