Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -19,6 +19,8 @@ processor = AutoProcessor.from_pretrained(model_id)
|
|
| 19 |
model = AutoModelForImageTextToText.from_pretrained(
|
| 20 |
model_id, device_map="auto", torch_dtype=torch.bfloat16
|
| 21 |
)
|
|
|
|
|
|
|
| 22 |
|
| 23 |
MAX_NUM_IMAGES = int(os.getenv("MAX_NUM_IMAGES", "5"))
|
| 24 |
|
|
@@ -188,6 +190,7 @@ def run(message: dict, history: list[dict], system_prompt: str = "", max_new_tok
|
|
| 188 |
inputs,
|
| 189 |
max_new_tokens=max_new_tokens,
|
| 190 |
streamer=streamer,
|
|
|
|
| 191 |
temperature=1.0,
|
| 192 |
top_p=0.95,
|
| 193 |
top_k=64,
|
|
@@ -217,7 +220,7 @@ demo = gr.ChatInterface(
|
|
| 217 |
gr.Textbox(label="System Prompt", value="You are a helpful medical expert."),
|
| 218 |
gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
|
| 219 |
],
|
| 220 |
-
stop_btn=
|
| 221 |
title="MedGemma 4B IT",
|
| 222 |
description=DESCRIPTION,
|
| 223 |
run_examples_on_click=False,
|
|
|
|
| 19 |
model = AutoModelForImageTextToText.from_pretrained(
|
| 20 |
model_id, device_map="auto", torch_dtype=torch.bfloat16
|
| 21 |
)
|
| 22 |
+
# Update the model's generation config to enable sampling by default
|
| 23 |
+
model.generation_config.do_sample = True
|
| 24 |
|
| 25 |
MAX_NUM_IMAGES = int(os.getenv("MAX_NUM_IMAGES", "5"))
|
| 26 |
|
|
|
|
| 190 |
inputs,
|
| 191 |
max_new_tokens=max_new_tokens,
|
| 192 |
streamer=streamer,
|
| 193 |
+
do_sample=True,
|
| 194 |
temperature=1.0,
|
| 195 |
top_p=0.95,
|
| 196 |
top_k=64,
|
|
|
|
| 220 |
gr.Textbox(label="System Prompt", value="You are a helpful medical expert."),
|
| 221 |
gr.Slider(label="Max New Tokens", minimum=100, maximum=8192, step=10, value=2048),
|
| 222 |
],
|
| 223 |
+
stop_btn=True,
|
| 224 |
title="MedGemma 4B IT",
|
| 225 |
description=DESCRIPTION,
|
| 226 |
run_examples_on_click=False,
|