Spaces:
Running
Running
Commit
·
ef219f6
1
Parent(s):
8c8e27f
Upload 2 files
Browse files- app.py +7 -2
- requirements.txt +2 -1
app.py
CHANGED
|
@@ -1,16 +1,17 @@
|
|
| 1 |
-
import random
|
| 2 |
import time
|
| 3 |
import os
|
| 4 |
import gradio as gr
|
| 5 |
from text_generation import Client
|
| 6 |
from conversation import get_default_conv_template
|
|
|
|
| 7 |
|
| 8 |
|
| 9 |
endpoint_url = os.environ.get("ENDPOINT_URL", "http://127.0.0.1:8080")
|
| 10 |
client = Client(endpoint_url, timeout=120)
|
| 11 |
eos_token = "</s>"
|
|
|
|
| 12 |
|
| 13 |
-
|
| 14 |
|
| 15 |
with gr.Blocks() as demo:
|
| 16 |
chatbot = gr.Chatbot()
|
|
@@ -27,6 +28,10 @@ with gr.Blocks() as demo:
|
|
| 27 |
conv.append_message(roles['human'], user)
|
| 28 |
conv.append_message(roles["gpt"], bot)
|
| 29 |
msg = conv.get_prompt()
|
|
|
|
|
|
|
|
|
|
|
|
|
| 30 |
|
| 31 |
history[-1][1] = ""
|
| 32 |
for response in client.generate_stream(
|
|
|
|
|
|
|
| 1 |
import time
|
| 2 |
import os
|
| 3 |
import gradio as gr
|
| 4 |
from text_generation import Client
|
| 5 |
from conversation import get_default_conv_template
|
| 6 |
+
from transformers import AutoTokenizer
|
| 7 |
|
| 8 |
|
| 9 |
endpoint_url = os.environ.get("ENDPOINT_URL", "http://127.0.0.1:8080")
|
| 10 |
client = Client(endpoint_url, timeout=120)
|
| 11 |
eos_token = "</s>"
|
| 12 |
+
max_prompt_length = 4000
|
| 13 |
|
| 14 |
+
tokenizer = AutoTokenizer.from_pretrained("yentinglin/Taiwan-LLaMa-v1.0")
|
| 15 |
|
| 16 |
with gr.Blocks() as demo:
|
| 17 |
chatbot = gr.Chatbot()
|
|
|
|
| 28 |
conv.append_message(roles['human'], user)
|
| 29 |
conv.append_message(roles["gpt"], bot)
|
| 30 |
msg = conv.get_prompt()
|
| 31 |
+
prompt_tokens = tokenizer.encode(msg)
|
| 32 |
+
length_of_prompt = len(prompt_tokens)
|
| 33 |
+
if length_of_prompt > max_prompt_length:
|
| 34 |
+
msg = tokenizer.decode(prompt_tokens[-max_prompt_length:])
|
| 35 |
|
| 36 |
history[-1][1] = ""
|
| 37 |
for response in client.generate_stream(
|
requirements.txt
CHANGED
|
@@ -1 +1,2 @@
|
|
| 1 |
-
text-generation==0.6.0
|
|
|
|
|
|
| 1 |
+
text-generation==0.6.0
|
| 2 |
+
transformers==4.31.0
|