Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -513,16 +513,14 @@ def get_zip_download_link(zip_file):
|
|
| 513 |
|
| 514 |
# 14. Inference Endpoints for Whisper (best fastest STT) on NVIDIA T4 and Llama (best fastest AGI LLM) on NVIDIA A10
|
| 515 |
# My Inference Endpoint
|
| 516 |
-
API_URL_IE = f'https://tonpixzfvq3791u9.us-east-1.aws.endpoints.huggingface.cloud'
|
| 517 |
# Original
|
| 518 |
-
API_URL_IE = "https://api-inference.huggingface.co/models/openai/whisper-small.en"
|
|
|
|
|
|
|
| 519 |
MODEL2 = "openai/whisper-small.en"
|
| 520 |
MODEL2_URL = "https://huggingface.co/openai/whisper-small.en"
|
| 521 |
-
|
| 522 |
-
# "Authorization": "Bearer XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX",
|
| 523 |
-
# "Content-Type": "audio/wav"
|
| 524 |
-
#}
|
| 525 |
-
# HF_KEY = os.getenv('HF_KEY')
|
| 526 |
HF_KEY = st.secrets['HF_KEY']
|
| 527 |
headers = {
|
| 528 |
"Authorization": f"Bearer {HF_KEY}",
|
|
|
|
| 513 |
|
| 514 |
# 14. Inference Endpoints for Whisper (best fastest STT) on NVIDIA T4 and Llama (best fastest AGI LLM) on NVIDIA A10
|
| 515 |
# My Inference Endpoint
|
| 516 |
+
# API_URL_IE = f'https://tonpixzfvq3791u9.us-east-1.aws.endpoints.huggingface.cloud'
|
| 517 |
# Original
|
| 518 |
+
#API_URL_IE = "https://api-inference.huggingface.co/models/openai/whisper-small.en"
|
| 519 |
+
# Latest small - 1/26
|
| 520 |
+
API_URL_IE = "https://api-inference.huggingface.co/models/openai/whisper-small"
|
| 521 |
MODEL2 = "openai/whisper-small.en"
|
| 522 |
MODEL2_URL = "https://huggingface.co/openai/whisper-small.en"
|
| 523 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
| 524 |
HF_KEY = st.secrets['HF_KEY']
|
| 525 |
headers = {
|
| 526 |
"Authorization": f"Bearer {HF_KEY}",
|