Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
 - added_tokens.json +5 -0
 - config.json +46 -0
 - generation_config.json +6 -0
 - merges.txt +0 -0
 - onnx/decoder_model_merged.onnx +3 -0
 - onnx/decoder_model_merged_bnb4.onnx +3 -0
 - onnx/decoder_model_merged_fp16.onnx +3 -0
 - onnx/decoder_model_merged_int8.onnx +3 -0
 - onnx/decoder_model_merged_q4.onnx +3 -0
 - onnx/decoder_model_merged_q4f16.onnx +3 -0
 - onnx/decoder_model_merged_quantized.onnx +3 -0
 - onnx/decoder_model_merged_uint8.onnx +3 -0
 - special_tokens_map.json +20 -0
 - tokenizer.json +3 -0
 - tokenizer_config.json +45 -0
 - vocab.json +0 -0
 
    	
        .gitattributes
    CHANGED
    
    | 
         @@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text 
     | 
|
| 33 | 
         
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         
     | 
| 34 | 
         
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         
     | 
| 35 | 
         
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         
     | 
| 
         | 
| 
         | 
|
| 33 | 
         
             
            *.zip filter=lfs diff=lfs merge=lfs -text
         
     | 
| 34 | 
         
             
            *.zst filter=lfs diff=lfs merge=lfs -text
         
     | 
| 35 | 
         
             
            *tfevents* filter=lfs diff=lfs merge=lfs -text
         
     | 
| 36 | 
         
            +
            tokenizer.json filter=lfs diff=lfs merge=lfs -text
         
     | 
    	
        added_tokens.json
    ADDED
    
    | 
         @@ -0,0 +1,5 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "<|endoftext|>": 151643,
         
     | 
| 3 | 
         
            +
              "<|im_end|>": 151645,
         
     | 
| 4 | 
         
            +
              "<|im_start|>": 151644
         
     | 
| 5 | 
         
            +
            }
         
     | 
    	
        config.json
    ADDED
    
    | 
         @@ -0,0 +1,46 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_name_or_path": "./qwen2-lm",
         
     | 
| 3 | 
         
            +
              "architectures": [
         
     | 
| 4 | 
         
            +
                "Qwen2ForCausalLM"
         
     | 
| 5 | 
         
            +
              ],
         
     | 
| 6 | 
         
            +
              "attention_dropout": 0.0,
         
     | 
| 7 | 
         
            +
              "bos_token_id": 151643,
         
     | 
| 8 | 
         
            +
              "eos_token_id": 151645,
         
     | 
| 9 | 
         
            +
              "freeze_mm_mlp_adapter": false,
         
     | 
| 10 | 
         
            +
              "hidden_act": "silu",
         
     | 
| 11 | 
         
            +
              "hidden_size": 896,
         
     | 
| 12 | 
         
            +
              "image_aspect_ratio": "pad",
         
     | 
| 13 | 
         
            +
              "image_grid_pinpoints": null,
         
     | 
| 14 | 
         
            +
              "initializer_range": 0.02,
         
     | 
| 15 | 
         
            +
              "intermediate_size": 4864,
         
     | 
| 16 | 
         
            +
              "max_position_embeddings": 32768,
         
     | 
| 17 | 
         
            +
              "max_window_layers": 24,
         
     | 
| 18 | 
         
            +
              "mm_hidden_size": 3072,
         
     | 
| 19 | 
         
            +
              "mm_patch_merge_type": "flat",
         
     | 
| 20 | 
         
            +
              "mm_projector_lr": null,
         
     | 
| 21 | 
         
            +
              "mm_projector_type": "mlp2x_gelu",
         
     | 
| 22 | 
         
            +
              "mm_use_im_patch_token": false,
         
     | 
| 23 | 
         
            +
              "mm_use_im_start_end": false,
         
     | 
| 24 | 
         
            +
              "mm_vision_select_feature": "patch",
         
     | 
| 25 | 
         
            +
              "mm_vision_select_layer": -2,
         
     | 
| 26 | 
         
            +
              "mm_vision_tower": "mobileclip_l_1024",
         
     | 
| 27 | 
         
            +
              "model_type": "qwen2",
         
     | 
| 28 | 
         
            +
              "num_attention_heads": 14,
         
     | 
| 29 | 
         
            +
              "num_hidden_layers": 24,
         
     | 
| 30 | 
         
            +
              "num_key_value_heads": 2,
         
     | 
| 31 | 
         
            +
              "rms_norm_eps": 1e-06,
         
     | 
| 32 | 
         
            +
              "rope_scaling": null,
         
     | 
| 33 | 
         
            +
              "rope_theta": 1000000.0,
         
     | 
| 34 | 
         
            +
              "sliding_window": null,
         
     | 
| 35 | 
         
            +
              "tie_word_embeddings": true,
         
     | 
| 36 | 
         
            +
              "tokenizer_model_max_length": 8192,
         
     | 
| 37 | 
         
            +
              "tokenizer_padding_side": "right",
         
     | 
| 38 | 
         
            +
              "torch_dtype": "float32",
         
     | 
| 39 | 
         
            +
              "transformers_version": "4.48.3",
         
     | 
| 40 | 
         
            +
              "tune_mm_mlp_adapter": false,
         
     | 
| 41 | 
         
            +
              "unfreeze_mm_vision_tower": true,
         
     | 
| 42 | 
         
            +
              "use_cache": true,
         
     | 
| 43 | 
         
            +
              "use_mm_proj": true,
         
     | 
| 44 | 
         
            +
              "use_sliding_window": false,
         
     | 
| 45 | 
         
            +
              "vocab_size": 151646
         
     | 
| 46 | 
         
            +
            }
         
     | 
    	
        generation_config.json
    ADDED
    
    | 
         @@ -0,0 +1,6 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "_from_model_config": true,
         
     | 
| 3 | 
         
            +
              "bos_token_id": 151643,
         
     | 
| 4 | 
         
            +
              "eos_token_id": 151645,
         
     | 
| 5 | 
         
            +
              "transformers_version": "4.48.3"
         
     | 
| 6 | 
         
            +
            }
         
     | 
    	
        merges.txt
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         | 
    	
        onnx/decoder_model_merged.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:9aeeb8ab6653e79f5f6d6bec7c59c8faea183014dcae1e922514e08c084b9ed4
         
     | 
| 3 | 
         
            +
            size 1983654988
         
     | 
    	
        onnx/decoder_model_merged_bnb4.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:3249dd716dc77ef35a3fb6036b7530424c162cff8f3da5b6615844b65cf2a490
         
     | 
| 3 | 
         
            +
            size 286590783
         
     | 
    	
        onnx/decoder_model_merged_fp16.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:0b79527daab2ac1566e1a5e3947c0839e2aa5e82ee2cce2d5bb1c6b7f3d27b21
         
     | 
| 3 | 
         
            +
            size 991914846
         
     | 
    	
        onnx/decoder_model_merged_int8.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:93ee4552127e1c12549b0b931fb6cf90e0d794faedfe5e56b8b5b90f04e63979
         
     | 
| 3 | 
         
            +
            size 502744473
         
     | 
    	
        onnx/decoder_model_merged_q4.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:d4efcd943ec6bc3e0ddbcd2022330a563243ec9a648fb05a616b201d5a8c43a5
         
     | 
| 3 | 
         
            +
            size 317445767
         
     | 
    	
        onnx/decoder_model_merged_q4f16.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:6ea00b526e59a5087e90e8e73b74a09347a7f1127f052476211b03aecca3fb0d
         
     | 
| 3 | 
         
            +
            size 282252137
         
     | 
    	
        onnx/decoder_model_merged_quantized.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:0fd14a3135022c608ec9d4c049509759bd56d0cc621858c3fffa6e5b765627c1
         
     | 
| 3 | 
         
            +
            size 502744556
         
     | 
    	
        onnx/decoder_model_merged_uint8.onnx
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:0fd14a3135022c608ec9d4c049509759bd56d0cc621858c3fffa6e5b765627c1
         
     | 
| 3 | 
         
            +
            size 502744556
         
     | 
    	
        special_tokens_map.json
    ADDED
    
    | 
         @@ -0,0 +1,20 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "additional_special_tokens": [
         
     | 
| 3 | 
         
            +
                "<|im_start|>",
         
     | 
| 4 | 
         
            +
                "<|im_end|>"
         
     | 
| 5 | 
         
            +
              ],
         
     | 
| 6 | 
         
            +
              "eos_token": {
         
     | 
| 7 | 
         
            +
                "content": "<|im_end|>",
         
     | 
| 8 | 
         
            +
                "lstrip": false,
         
     | 
| 9 | 
         
            +
                "normalized": false,
         
     | 
| 10 | 
         
            +
                "rstrip": false,
         
     | 
| 11 | 
         
            +
                "single_word": false
         
     | 
| 12 | 
         
            +
              },
         
     | 
| 13 | 
         
            +
              "pad_token": {
         
     | 
| 14 | 
         
            +
                "content": "<|endoftext|>",
         
     | 
| 15 | 
         
            +
                "lstrip": false,
         
     | 
| 16 | 
         
            +
                "normalized": false,
         
     | 
| 17 | 
         
            +
                "rstrip": false,
         
     | 
| 18 | 
         
            +
                "single_word": false
         
     | 
| 19 | 
         
            +
              }
         
     | 
| 20 | 
         
            +
            }
         
     | 
    	
        tokenizer.json
    ADDED
    
    | 
         @@ -0,0 +1,3 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            version https://git-lfs.github.com/spec/v1
         
     | 
| 2 | 
         
            +
            oid sha256:22a32bc7af1fc17ed370988966140a379f0421b0a508e4ae3fe4bcf7a86644e1
         
     | 
| 3 | 
         
            +
            size 11413100
         
     | 
    	
        tokenizer_config.json
    ADDED
    
    | 
         @@ -0,0 +1,45 @@ 
     | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
|
| 
         | 
| 
         | 
|
| 1 | 
         
            +
            {
         
     | 
| 2 | 
         
            +
              "add_prefix_space": false,
         
     | 
| 3 | 
         
            +
              "added_tokens_decoder": {
         
     | 
| 4 | 
         
            +
                "151643": {
         
     | 
| 5 | 
         
            +
                  "content": "<|endoftext|>",
         
     | 
| 6 | 
         
            +
                  "lstrip": false,
         
     | 
| 7 | 
         
            +
                  "normalized": false,
         
     | 
| 8 | 
         
            +
                  "rstrip": false,
         
     | 
| 9 | 
         
            +
                  "single_word": false,
         
     | 
| 10 | 
         
            +
                  "special": true
         
     | 
| 11 | 
         
            +
                },
         
     | 
| 12 | 
         
            +
                "151644": {
         
     | 
| 13 | 
         
            +
                  "content": "<|im_start|>",
         
     | 
| 14 | 
         
            +
                  "lstrip": false,
         
     | 
| 15 | 
         
            +
                  "normalized": false,
         
     | 
| 16 | 
         
            +
                  "rstrip": false,
         
     | 
| 17 | 
         
            +
                  "single_word": false,
         
     | 
| 18 | 
         
            +
                  "special": true
         
     | 
| 19 | 
         
            +
                },
         
     | 
| 20 | 
         
            +
                "151645": {
         
     | 
| 21 | 
         
            +
                  "content": "<|im_end|>",
         
     | 
| 22 | 
         
            +
                  "lstrip": false,
         
     | 
| 23 | 
         
            +
                  "normalized": false,
         
     | 
| 24 | 
         
            +
                  "rstrip": false,
         
     | 
| 25 | 
         
            +
                  "single_word": false,
         
     | 
| 26 | 
         
            +
                  "special": true
         
     | 
| 27 | 
         
            +
                }
         
     | 
| 28 | 
         
            +
              },
         
     | 
| 29 | 
         
            +
              "additional_special_tokens": [
         
     | 
| 30 | 
         
            +
                "<|im_start|>",
         
     | 
| 31 | 
         
            +
                "<|im_end|>"
         
     | 
| 32 | 
         
            +
              ],
         
     | 
| 33 | 
         
            +
              "bos_token": null,
         
     | 
| 34 | 
         
            +
              "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
         
     | 
| 35 | 
         
            +
              "clean_up_tokenization_spaces": false,
         
     | 
| 36 | 
         
            +
              "eos_token": "<|im_end|>",
         
     | 
| 37 | 
         
            +
              "errors": "replace",
         
     | 
| 38 | 
         
            +
              "extra_special_tokens": {},
         
     | 
| 39 | 
         
            +
              "model_max_length": 8192,
         
     | 
| 40 | 
         
            +
              "pad_token": "<|endoftext|>",
         
     | 
| 41 | 
         
            +
              "padding_side": "right",
         
     | 
| 42 | 
         
            +
              "split_special_tokens": false,
         
     | 
| 43 | 
         
            +
              "tokenizer_class": "Qwen2Tokenizer",
         
     | 
| 44 | 
         
            +
              "unk_token": null
         
     | 
| 45 | 
         
            +
            }
         
     | 
    	
        vocab.json
    ADDED
    
    | 
         The diff for this file is too large to render. 
		See raw diff 
     | 
| 
         |