Spaces:
Paused
Paused
| from transformers import AutoTokenizer | |
| import os | |
| class Tokenizer: | |
| def __init__(self) -> None: | |
| self.tokenizer = AutoTokenizer.from_pretrained("openai-community/gpt2", hf_token = '...') | |
| self.tokenizer.add_special_tokens({'pad_token': '[PAD]'}) | |
| def ready_tokenizer(self): | |
| return self.tokenizer | |