manueljohnson063 commited on
Commit
7f2cf1a
·
verified ·
1 Parent(s): 3d2fc68

Upload app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -7
app.py CHANGED
@@ -13,9 +13,6 @@ model_id = "nvidia/canary-qwen-2.5b"
13
  print("Loading NVIDIA Canary-Qwen-2.5B model using NeMo...")
14
  model = SALM.from_pretrained(model_id)
15
 
16
- device = "cuda" if torch.cuda.is_available() else "cpu"
17
- model = model.to(device)
18
-
19
  def generate_text(prompt, max_tokens=200, temperature=0.7, top_p=0.9):
20
  """Generate text using the NVIDIA NeMo model (LLM mode)"""
21
 
@@ -31,8 +28,7 @@ def generate_text(prompt, max_tokens=200, temperature=0.7, top_p=0.9):
31
  )
32
 
33
  # Convert IDs to text using model's tokenizer
34
- # response = model.tokenizer.ids_to_text(answer_ids[0].cpu())
35
- response = model.tokenizer.ids_to_text(answer_ids[0].to(device))
36
  return response
37
 
38
  except Exception as e:
@@ -54,8 +50,7 @@ def transcribe_audio(audio_file, user_prompt="Transcribe the following:"):
54
  )
55
 
56
  # Convert IDs to text
57
- # transcript = model.tokenizer.ids_to_text(answer_ids[0].cpu())
58
- transcript = model.tokenizer.ids_to_text(answer_ids[0].to(device))
59
  return transcript
60
 
61
  except Exception as e:
@@ -238,4 +233,5 @@ with gr.Blocks(title="NVIDIA Canary-Qwen-2.5B Chat") as demo:
238
  )
239
 
240
  if __name__ == "__main__":
 
241
  demo.launch(share=True)
 
13
  print("Loading NVIDIA Canary-Qwen-2.5B model using NeMo...")
14
  model = SALM.from_pretrained(model_id)
15
 
 
 
 
16
  def generate_text(prompt, max_tokens=200, temperature=0.7, top_p=0.9):
17
  """Generate text using the NVIDIA NeMo model (LLM mode)"""
18
 
 
28
  )
29
 
30
  # Convert IDs to text using model's tokenizer
31
+ response = model.tokenizer.ids_to_text(answer_ids[0].cpu())
 
32
  return response
33
 
34
  except Exception as e:
 
50
  )
51
 
52
  # Convert IDs to text
53
+ transcript = model.tokenizer.ids_to_text(answer_ids[0].cpu())
 
54
  return transcript
55
 
56
  except Exception as e:
 
233
  )
234
 
235
  if __name__ == "__main__":
236
+ demo.queue(api_open=True)
237
  demo.launch(share=True)