Spaces:
Runtime error
Runtime error
| import streamlit as st | |
| from transformers import GPT2LMHeadModel, GPT2Tokenizer | |
| import torch | |
| DEVICE = torch.device('cuda' if torch.cuda.is_available() else 'cpu') | |
| model_path = "/zhvanetsky_model" | |
| tokenizer = GPT2Tokenizer.from_pretrained(model_path) | |
| model = GPT2LMHeadModel.from_pretrained(model_path).to(DEVICE) | |
| def generate_text(input_text): | |
| model.eval() | |
| input_ids = tokenizer.encode(input_text, return_tensors="pt").to(DEVICE) | |
| with torch.no_grad(): | |
| out = model.generate(input_ids, | |
| do_sample=True, | |
| num_beams=10, | |
| temperature=2.2, | |
| top_p=0.85, | |
| top_k=500, | |
| max_length=100, | |
| no_repeat_ngram_size=3, | |
| num_return_sequences=3, | |
| ) | |
| return tokenizer.decode(out[0], skip_special_tokens=True) | |
| st.title("GPT-2 Text Generator") | |
| user_input = st.text_area("Input Text", "Введите ваш текст") | |
| if st.button("Generate"): | |
| generated_output = generate_text(user_input) | |
| st.text_area("Generated Text", generated_output) |