Spaces:
Running
on
Zero
Running
on
Zero
Add share option
Browse files
app.py
CHANGED
|
@@ -10,6 +10,7 @@ from huggingface_hub import model_info
|
|
| 10 |
MODEL_NAME = "NbAiLab/whisper-large-sme" #this always needs to stay in line 8 :D sorry for the hackiness
|
| 11 |
lang = "fi"
|
| 12 |
|
|
|
|
| 13 |
auth_token = os.environ.get("AUTH_TOKEN") or True
|
| 14 |
device = 0 if torch.cuda.is_available() else "cpu"
|
| 15 |
pipe = pipeline(
|
|
@@ -99,4 +100,4 @@ yt_transcribe = gr.Interface(
|
|
| 99 |
with demo:
|
| 100 |
gr.TabbedInterface([mf_transcribe, yt_transcribe], ["Transcribe Audio", "Transcribe YouTube"])
|
| 101 |
|
| 102 |
-
demo.launch(enable_queue=True)
|
|
|
|
| 10 |
MODEL_NAME = "NbAiLab/whisper-large-sme" #this always needs to stay in line 8 :D sorry for the hackiness
|
| 11 |
lang = "fi"
|
| 12 |
|
| 13 |
+
share = (os.environ.get("SHARE", "False")[0].lower() in "ty1") or None
|
| 14 |
auth_token = os.environ.get("AUTH_TOKEN") or True
|
| 15 |
device = 0 if torch.cuda.is_available() else "cpu"
|
| 16 |
pipe = pipeline(
|
|
|
|
| 100 |
with demo:
|
| 101 |
gr.TabbedInterface([mf_transcribe, yt_transcribe], ["Transcribe Audio", "Transcribe YouTube"])
|
| 102 |
|
| 103 |
+
demo.launch(enable_queue=True, share=True)
|