Update tokenizer_config.json
Browse files- tokenizer_config.json +15 -15
    	
        tokenizer_config.json
    CHANGED
    
    | @@ -1,17 +1,17 @@ | |
| 1 | 
             
            {
         | 
| 2 | 
            -
            "add_bos_token": false,
         | 
| 3 | 
            -
            "add_eos_token": false,
         | 
| 4 | 
            -
            "bos_token": "<|startoftext|>",
         | 
| 5 | 
            -
            "chat_template": "{ | 
| 6 | 
            -
            "clean_up_tokenization_spaces": false,
         | 
| 7 | 
            -
            "cls_token": "[CLS]",
         | 
| 8 | 
            -
            "eos_token": "<|endoftext|>",
         | 
| 9 | 
            -
            "fast_tokenizer": true,
         | 
| 10 | 
            -
            "gmask_token": "[gMASK]",
         | 
| 11 | 
            -
            "merges_file": null,
         | 
| 12 | 
            -
            "model_max_length": 1000000000000000019884624838656,
         | 
| 13 | 
            -
            "pad_token": "<|endoftext|>",
         | 
| 14 | 
            -
            "tokenizer_class": "PreTrainedTokenizerFast",
         | 
| 15 | 
            -
            "trust_remote_code": true,
         | 
| 16 | 
            -
            "vocab_file": null
         | 
| 17 | 
             
            }
         | 
|  | |
| 1 | 
             
            {
         | 
| 2 | 
            +
              "add_bos_token": false,
         | 
| 3 | 
            +
              "add_eos_token": false,
         | 
| 4 | 
            +
              "bos_token": "<|startoftext|>",
         | 
| 5 | 
            +
              "chat_template": "{% for message in messages %}{% set role = message['role'] | lower %}{% if role == 'user' %}{% set role = 'HUMAN' %}{% endif %}{% set role = role | upper %}{{ '<role>' + role + '</role>' + message['content'] }}{% endfor %}{% if add_generation_prompt %}{{ '<role>ASSISTANT</role>' }}{% endif %}",
         | 
| 6 | 
            +
              "clean_up_tokenization_spaces": false,
         | 
| 7 | 
            +
              "cls_token": "[CLS]",
         | 
| 8 | 
            +
              "eos_token": "<|endoftext|>",
         | 
| 9 | 
            +
              "fast_tokenizer": true,
         | 
| 10 | 
            +
              "gmask_token": "[gMASK]",
         | 
| 11 | 
            +
              "merges_file": null,
         | 
| 12 | 
            +
              "model_max_length": 1000000000000000019884624838656,
         | 
| 13 | 
            +
              "pad_token": "<|endoftext|>",
         | 
| 14 | 
            +
              "tokenizer_class": "PreTrainedTokenizerFast",
         | 
| 15 | 
            +
              "trust_remote_code": true,
         | 
| 16 | 
            +
              "vocab_file": null
         | 
| 17 | 
             
            }
         | 
