Update app.py
Browse files
app.py
CHANGED
|
@@ -55,7 +55,6 @@ def load_models():
|
|
| 55 |
tokenizer = AutoTokenizer.from_pretrained('gpt2-large')
|
| 56 |
model = GPT2LMHeadModel.from_pretrained('gpt2-large')
|
| 57 |
return tokenizer, model
|
| 58 |
-
tokenizer, model = load_models()
|
| 59 |
# Function For Extracting Entities
|
| 60 |
@st.experimental_singleton
|
| 61 |
def entity_analyzer(my_text):
|
|
@@ -103,6 +102,7 @@ def main():
|
|
| 103 |
message = st.text_area("Enter the Text","Type please ..")
|
| 104 |
input_ids = tokenizer(message, return_tensors='pt').input_ids
|
| 105 |
if st.button("Generate"):
|
|
|
|
| 106 |
st.text("Using Hugging Face Trnsformer, Contrastive Search ..")
|
| 107 |
output = model.generate(input_ids, max_length=128)
|
| 108 |
st.success(tokenizer.decode(output[0], skip_special_tokens=True))
|
|
|
|
| 55 |
tokenizer = AutoTokenizer.from_pretrained('gpt2-large')
|
| 56 |
model = GPT2LMHeadModel.from_pretrained('gpt2-large')
|
| 57 |
return tokenizer, model
|
|
|
|
| 58 |
# Function For Extracting Entities
|
| 59 |
@st.experimental_singleton
|
| 60 |
def entity_analyzer(my_text):
|
|
|
|
| 102 |
message = st.text_area("Enter the Text","Type please ..")
|
| 103 |
input_ids = tokenizer(message, return_tensors='pt').input_ids
|
| 104 |
if st.button("Generate"):
|
| 105 |
+
tokenizer, model = load_models()
|
| 106 |
st.text("Using Hugging Face Trnsformer, Contrastive Search ..")
|
| 107 |
output = model.generate(input_ids, max_length=128)
|
| 108 |
st.success(tokenizer.decode(output[0], skip_special_tokens=True))
|