Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
Commit
·
78e9bc6
1
Parent(s):
b9392b7
New API models
Browse files- README.md +38 -0
- external_models_results.json +158 -3
- model_list.txt +38 -0
- update_models_in_readme.py +6 -2
README.md
CHANGED
|
@@ -113,6 +113,7 @@ models:
|
|
| 113 |
- EleutherAI/gpt-neo-125m
|
| 114 |
- EleutherAI/gpt-neo-2.7B
|
| 115 |
- EleutherAI/gpt-neox-20b
|
|
|
|
| 116 |
- EleutherAI/pythia-12b
|
| 117 |
- EleutherAI/pythia-14m
|
| 118 |
- EleutherAI/pythia-160m
|
|
@@ -135,8 +136,11 @@ models:
|
|
| 135 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
| 136 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
| 137 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
|
|
|
| 138 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
| 139 |
- Eurdem/Defne-llama3.1-8B
|
|
|
|
|
|
|
| 140 |
- FuseAI/FuseChat-7B-VaRM
|
| 141 |
- FuseAI/OpenChat-3.5-7B-Solar
|
| 142 |
- GeneZC/MiniChat-1.5-3B
|
|
@@ -201,6 +205,7 @@ models:
|
|
| 201 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
| 202 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
| 203 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
|
|
|
| 204 |
- MagusCorp/legislinho
|
| 205 |
- MarinaraSpaghetti/NemoReRemix-12B
|
| 206 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
@@ -212,6 +217,10 @@ models:
|
|
| 212 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
| 213 |
- MaziyarPanahi/Mistral-7B-v0.3
|
| 214 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
|
|
|
|
|
|
|
|
|
|
|
|
| 215 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
| 216 |
- MulaBR/Mula-4x160-v0.1
|
| 217 |
- MulaBR/Mula-8x160-v0.1
|
|
@@ -221,6 +230,8 @@ models:
|
|
| 221 |
- NTQAI/Nxcode-CQ-7B-orpo
|
| 222 |
- Nexusflow/Starling-LM-7B-beta
|
| 223 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
|
|
|
|
|
|
| 224 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
| 225 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
| 226 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
@@ -293,6 +304,7 @@ models:
|
|
| 293 |
- Qwen/Qwen2.5-1.5B-Instruct
|
| 294 |
- Qwen/Qwen2.5-14B
|
| 295 |
- Qwen/Qwen2.5-14B-Instruct
|
|
|
|
| 296 |
- Qwen/Qwen2.5-32B
|
| 297 |
- Qwen/Qwen2.5-32B-Instruct
|
| 298 |
- Qwen/Qwen2.5-3B
|
|
@@ -307,8 +319,13 @@ models:
|
|
| 307 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
| 308 |
- Qwen/Qwen2.5-Coder-7B
|
| 309 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
| 310 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
| 311 |
- Ramikan-BR/Qwen2-0.5B-v5
|
|
|
|
| 312 |
- Ramikan-BR/TiamaPY-v30
|
| 313 |
- Ramikan-BR/TiamaPY-v31
|
| 314 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
@@ -320,7 +337,9 @@ models:
|
|
| 320 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
| 321 |
- Sakalti/Oxyge1-33B
|
| 322 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
|
|
|
| 323 |
- Sao10K/L3-8B-Niitama-v1
|
|
|
|
| 324 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
| 325 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
| 326 |
- SeaLLMs/SeaLLM-7B-v2
|
|
@@ -347,6 +366,7 @@ models:
|
|
| 347 |
- TheBloke/wizardLM-7B-HF
|
| 348 |
- TheBloke/zephyr-7B-beta-GPTQ
|
| 349 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
|
|
|
| 350 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
| 351 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 352 |
- TinyLlama/TinyLlama_v1.1
|
|
@@ -368,6 +388,7 @@ models:
|
|
| 368 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
| 369 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
| 370 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
|
|
|
| 371 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
| 372 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
| 373 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
@@ -397,9 +418,12 @@ models:
|
|
| 397 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
| 398 |
- abacusai/Liberated-Qwen1.5-14B
|
| 399 |
- abacusai/Llama-3-Smaug-8B
|
|
|
|
| 400 |
- abacusai/Smaug-34B-v0.1
|
| 401 |
- abacusai/Smaug-72B-v0.1
|
| 402 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
|
|
|
|
|
|
| 403 |
- abhishek/autotrain-llama3-orpo-v2
|
| 404 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
| 405 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
@@ -445,6 +469,7 @@ models:
|
|
| 445 |
- arcee-ai/Arcee-Spark
|
| 446 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
| 447 |
- arcee-ai/SuperNova-Medius
|
|
|
|
| 448 |
- arcee-ai/Virtuoso-Small
|
| 449 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
| 450 |
- argilla/notus-7b-v1
|
|
@@ -472,6 +497,7 @@ models:
|
|
| 472 |
- botbot-ai/CabraLlama3-8b
|
| 473 |
- botbot-ai/CabraMistral-v3-7b-32k
|
| 474 |
- botbot-ai/CabraMixtral-8x7b
|
|
|
|
| 475 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
| 476 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
| 477 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
@@ -489,6 +515,10 @@ models:
|
|
| 489 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
| 490 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
| 491 |
- cnmoro/Mistral-7B-Portuguese
|
|
|
|
|
|
|
|
|
|
|
|
|
| 492 |
- cognitivecomputations/WestLake-7B-v2-laser
|
| 493 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
| 494 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
@@ -510,6 +540,7 @@ models:
|
|
| 510 |
- cognitivess/bella-1-8b
|
| 511 |
- cosmicvalor/mistral-orthogonalized
|
| 512 |
- croissantllm/CroissantLLMBase
|
|
|
|
| 513 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
| 514 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
| 515 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
@@ -573,6 +604,7 @@ models:
|
|
| 573 |
- freewheelin/free-solar-evo-v0.1
|
| 574 |
- freewheelin/free-solar-evo-v0.11
|
| 575 |
- freewheelin/free-solar-evo-v0.13
|
|
|
|
| 576 |
- ghost-x/ghost-8b-beta
|
| 577 |
- ghost-x/ghost-8b-beta-1608
|
| 578 |
- google/gemma-1.1-2b-it
|
|
@@ -733,6 +765,7 @@ models:
|
|
| 733 |
- microsoft/phi-1_5
|
| 734 |
- microsoft/phi-2
|
| 735 |
- microsoft/phi-4
|
|
|
|
| 736 |
- migtissera/Tess-M-v1.3
|
| 737 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
| 738 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
@@ -766,12 +799,14 @@ models:
|
|
| 766 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
| 767 |
- mlabonne/NeuralMonarch-7B
|
| 768 |
- monilouise/opt125M_portuguese
|
|
|
|
| 769 |
- mosaicml/mpt-30b
|
| 770 |
- mosaicml/mpt-7b
|
| 771 |
- mosaicml/mpt-7b-8k
|
| 772 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
| 773 |
- natong19/Qwen2-7B-Instruct-abliterated
|
| 774 |
- nbeerbower/gemma2-gutenberg-27B
|
|
|
|
| 775 |
- nicholasKluge/Aira-2-portuguese-124M
|
| 776 |
- nicholasKluge/Aira-2-portuguese-1B7
|
| 777 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
@@ -855,6 +890,7 @@ models:
|
|
| 855 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
| 856 |
- princeton-nlp/gemma-2-9b-it-DPO
|
| 857 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
|
|
|
| 858 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
| 859 |
- projecte-aina/FLOR-1.3B
|
| 860 |
- projecte-aina/FLOR-6.3B
|
|
@@ -933,11 +969,13 @@ models:
|
|
| 933 |
- shadowml/BeagSake-7B
|
| 934 |
- shadowml/Mixolar-4x7b
|
| 935 |
- sometimesanotion/Lamarck-14B-v0.6
|
|
|
|
| 936 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
| 937 |
- speakleash/Bielik-11B-v2
|
| 938 |
- speakleash/Bielik-11B-v2.0-Instruct
|
| 939 |
- speakleash/Bielik-11B-v2.1-Instruct
|
| 940 |
- speakleash/Bielik-11B-v2.2-Instruct
|
|
|
|
| 941 |
- ssmits/Falcon2-5.5B-Portuguese
|
| 942 |
- ssmits/Falcon2-5.5B-multilingual
|
| 943 |
- stabilityai/stablelm-2-12b
|
|
|
|
| 113 |
- EleutherAI/gpt-neo-125m
|
| 114 |
- EleutherAI/gpt-neo-2.7B
|
| 115 |
- EleutherAI/gpt-neox-20b
|
| 116 |
+
- EleutherAI/polyglot-ko-12.8b
|
| 117 |
- EleutherAI/pythia-12b
|
| 118 |
- EleutherAI/pythia-14m
|
| 119 |
- EleutherAI/pythia-160m
|
|
|
|
| 136 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
| 137 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
| 138 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
| 139 |
+
- EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
|
| 140 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
| 141 |
- Eurdem/Defne-llama3.1-8B
|
| 142 |
+
- FluxiIA/Tucaninho
|
| 143 |
+
- FluxiIA/Tucaninho_dpo
|
| 144 |
- FuseAI/FuseChat-7B-VaRM
|
| 145 |
- FuseAI/OpenChat-3.5-7B-Solar
|
| 146 |
- GeneZC/MiniChat-1.5-3B
|
|
|
|
| 205 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
| 206 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
| 207 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
| 208 |
+
- MagusCorp/grpo_lora_enem_llama3_7b
|
| 209 |
- MagusCorp/legislinho
|
| 210 |
- MarinaraSpaghetti/NemoReRemix-12B
|
| 211 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
|
|
| 217 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
| 218 |
- MaziyarPanahi/Mistral-7B-v0.3
|
| 219 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
| 220 |
+
- MaziyarPanahi/calme-2.1-qwen2-7b
|
| 221 |
+
- MaziyarPanahi/calme-2.2-qwen2-7b
|
| 222 |
+
- MaziyarPanahi/calme-2.3-qwen2-7b
|
| 223 |
+
- MaziyarPanahi/calme-2.4-qwen2-7b
|
| 224 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
| 225 |
- MulaBR/Mula-4x160-v0.1
|
| 226 |
- MulaBR/Mula-8x160-v0.1
|
|
|
|
| 230 |
- NTQAI/Nxcode-CQ-7B-orpo
|
| 231 |
- Nexusflow/Starling-LM-7B-beta
|
| 232 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
| 233 |
+
- Nos-PT/Llama-Carvalho-PT
|
| 234 |
+
- Nos-PT/Llama-Carvalho-PT-GL
|
| 235 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
| 236 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
| 237 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
|
|
| 304 |
- Qwen/Qwen2.5-1.5B-Instruct
|
| 305 |
- Qwen/Qwen2.5-14B
|
| 306 |
- Qwen/Qwen2.5-14B-Instruct
|
| 307 |
+
- Qwen/Qwen2.5-14B-Instruct-1M
|
| 308 |
- Qwen/Qwen2.5-32B
|
| 309 |
- Qwen/Qwen2.5-32B-Instruct
|
| 310 |
- Qwen/Qwen2.5-3B
|
|
|
|
| 319 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
| 320 |
- Qwen/Qwen2.5-Coder-7B
|
| 321 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
| 322 |
+
- Qwen/Qwen2.5-VL-32B-Instruct
|
| 323 |
+
- Qwen/Qwen2.5-VL-72B-Instruct
|
| 324 |
+
- RDson/WomboCombo-R1-Coder-14B-Preview
|
| 325 |
+
- RLHFlow/ArmoRM-Llama3-8B-v0.1
|
| 326 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
| 327 |
- Ramikan-BR/Qwen2-0.5B-v5
|
| 328 |
+
- Ramikan-BR/Qwen2-0.5B-v9
|
| 329 |
- Ramikan-BR/TiamaPY-v30
|
| 330 |
- Ramikan-BR/TiamaPY-v31
|
| 331 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
|
|
| 337 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
| 338 |
- Sakalti/Oxyge1-33B
|
| 339 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
| 340 |
+
- Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
|
| 341 |
- Sao10K/L3-8B-Niitama-v1
|
| 342 |
+
- Sao10K/L3.1-8B-Niitama-v1.1
|
| 343 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
| 344 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
| 345 |
- SeaLLMs/SeaLLM-7B-v2
|
|
|
|
| 366 |
- TheBloke/wizardLM-7B-HF
|
| 367 |
- TheBloke/zephyr-7B-beta-GPTQ
|
| 368 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
| 369 |
+
- TheDrummer/Gemmasutra-9B-v1
|
| 370 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
| 371 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 372 |
- TinyLlama/TinyLlama_v1.1
|
|
|
|
| 388 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
| 389 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
| 390 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
| 391 |
+
- VAGOsolutions/SauerkrautLM-Phi-3-medium
|
| 392 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
| 393 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
| 394 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
|
|
| 418 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
| 419 |
- abacusai/Liberated-Qwen1.5-14B
|
| 420 |
- abacusai/Llama-3-Smaug-8B
|
| 421 |
+
- abacusai/Slerp-CM-mist-dpo
|
| 422 |
- abacusai/Smaug-34B-v0.1
|
| 423 |
- abacusai/Smaug-72B-v0.1
|
| 424 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
| 425 |
+
- abacusai/bigstral-12b-32k
|
| 426 |
+
- abacusai/bigyi-15b
|
| 427 |
- abhishek/autotrain-llama3-orpo-v2
|
| 428 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
| 429 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
|
|
| 469 |
- arcee-ai/Arcee-Spark
|
| 470 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
| 471 |
- arcee-ai/SuperNova-Medius
|
| 472 |
+
- arcee-ai/Virtuoso-Lite
|
| 473 |
- arcee-ai/Virtuoso-Small
|
| 474 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
| 475 |
- argilla/notus-7b-v1
|
|
|
|
| 497 |
- botbot-ai/CabraLlama3-8b
|
| 498 |
- botbot-ai/CabraMistral-v3-7b-32k
|
| 499 |
- botbot-ai/CabraMixtral-8x7b
|
| 500 |
+
- bunnycore/HyperLlama-3.1-8B
|
| 501 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
| 502 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
| 503 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
|
|
| 515 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
| 516 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
| 517 |
- cnmoro/Mistral-7B-Portuguese
|
| 518 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
|
| 519 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v1
|
| 520 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v2
|
| 521 |
+
- cognitivecomputations/Dolphin3.0-R1-Mistral-24B
|
| 522 |
- cognitivecomputations/WestLake-7B-v2-laser
|
| 523 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
| 524 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
|
|
| 540 |
- cognitivess/bella-1-8b
|
| 541 |
- cosmicvalor/mistral-orthogonalized
|
| 542 |
- croissantllm/CroissantLLMBase
|
| 543 |
+
- deepseek-ai/DeepSeek-R1-Distill-Llama-70B
|
| 544 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
| 545 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
| 546 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
|
|
| 604 |
- freewheelin/free-solar-evo-v0.1
|
| 605 |
- freewheelin/free-solar-evo-v0.11
|
| 606 |
- freewheelin/free-solar-evo-v0.13
|
| 607 |
+
- gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
|
| 608 |
- ghost-x/ghost-8b-beta
|
| 609 |
- ghost-x/ghost-8b-beta-1608
|
| 610 |
- google/gemma-1.1-2b-it
|
|
|
|
| 765 |
- microsoft/phi-1_5
|
| 766 |
- microsoft/phi-2
|
| 767 |
- microsoft/phi-4
|
| 768 |
+
- migtissera/Tess-3-Mistral-Nemo-12B
|
| 769 |
- migtissera/Tess-M-v1.3
|
| 770 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
| 771 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
|
|
| 799 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
| 800 |
- mlabonne/NeuralMonarch-7B
|
| 801 |
- monilouise/opt125M_portuguese
|
| 802 |
+
- monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
|
| 803 |
- mosaicml/mpt-30b
|
| 804 |
- mosaicml/mpt-7b
|
| 805 |
- mosaicml/mpt-7b-8k
|
| 806 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
| 807 |
- natong19/Qwen2-7B-Instruct-abliterated
|
| 808 |
- nbeerbower/gemma2-gutenberg-27B
|
| 809 |
+
- nbeerbower/mistral-nemo-wissenschaft-12B
|
| 810 |
- nicholasKluge/Aira-2-portuguese-124M
|
| 811 |
- nicholasKluge/Aira-2-portuguese-1B7
|
| 812 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
|
|
| 890 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
| 891 |
- princeton-nlp/gemma-2-9b-it-DPO
|
| 892 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
| 893 |
+
- prithivMLmods/Megatron-Opus-14B-Exp
|
| 894 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
| 895 |
- projecte-aina/FLOR-1.3B
|
| 896 |
- projecte-aina/FLOR-6.3B
|
|
|
|
| 969 |
- shadowml/BeagSake-7B
|
| 970 |
- shadowml/Mixolar-4x7b
|
| 971 |
- sometimesanotion/Lamarck-14B-v0.6
|
| 972 |
+
- sometimesanotion/Lamarck-14B-v0.7
|
| 973 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
| 974 |
- speakleash/Bielik-11B-v2
|
| 975 |
- speakleash/Bielik-11B-v2.0-Instruct
|
| 976 |
- speakleash/Bielik-11B-v2.1-Instruct
|
| 977 |
- speakleash/Bielik-11B-v2.2-Instruct
|
| 978 |
+
- speakleash/Bielik-11B-v2.3-Instruct
|
| 979 |
- ssmits/Falcon2-5.5B-Portuguese
|
| 980 |
- ssmits/Falcon2-5.5B-multilingual
|
| 981 |
- stabilityai/stablelm-2-12b
|
external_models_results.json
CHANGED
|
@@ -176,8 +176,8 @@
|
|
| 176 |
"result_metrics_npm": 0.6834036936130392
|
| 177 |
},
|
| 178 |
{
|
| 179 |
-
"model": "gemini-1.5-flash",
|
| 180 |
-
"name": "Gemini 1.5 Flash",
|
| 181 |
"link": "https://cloud.google.com/vertex-ai",
|
| 182 |
"date": "2024-08-09",
|
| 183 |
"status": "full",
|
|
@@ -222,7 +222,7 @@
|
|
| 222 |
{
|
| 223 |
"model": "nemotron-4-340b-instruct",
|
| 224 |
"name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
|
| 225 |
-
"link": "https://
|
| 226 |
"date": "2024-06-30",
|
| 227 |
"status": "full",
|
| 228 |
"main_language": "English",
|
|
@@ -333,6 +333,51 @@
|
|
| 333 |
"result_metrics_average": 0.7928134532766066,
|
| 334 |
"result_metrics_npm": 0.6915070359785283
|
| 335 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 336 |
{
|
| 337 |
"model": "gemini-2.0-flash-001",
|
| 338 |
"name": "Gemini 2.0 Flash (001)",
|
|
@@ -376,5 +421,115 @@
|
|
| 376 |
},
|
| 377 |
"result_metrics_average": 0.8056048352614735,
|
| 378 |
"result_metrics_npm": 0.6986042497176748
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 379 |
}
|
| 380 |
]
|
|
|
|
| 176 |
"result_metrics_npm": 0.6834036936130392
|
| 177 |
},
|
| 178 |
{
|
| 179 |
+
"model": "gemini-1.5-flash-001",
|
| 180 |
+
"name": "Gemini 1.5 Flash (001)",
|
| 181 |
"link": "https://cloud.google.com/vertex-ai",
|
| 182 |
"date": "2024-08-09",
|
| 183 |
"status": "full",
|
|
|
|
| 222 |
{
|
| 223 |
"model": "nemotron-4-340b-instruct",
|
| 224 |
"name": "nvidia/Nemotron-4-340B-Instruct (Nvidia API)",
|
| 225 |
+
"link": "https://huggingface.co/nvidia/Nemotron-4-340B-Instruct",
|
| 226 |
"date": "2024-06-30",
|
| 227 |
"status": "full",
|
| 228 |
"main_language": "English",
|
|
|
|
| 333 |
"result_metrics_average": 0.7928134532766066,
|
| 334 |
"result_metrics_npm": 0.6915070359785283
|
| 335 |
},
|
| 336 |
+
{
|
| 337 |
+
"model": "gemini-1.5-flash-002",
|
| 338 |
+
"name": "Gemini 1.5 Flash (002)",
|
| 339 |
+
"link": "https://cloud.google.com/vertex-ai",
|
| 340 |
+
"date": "2025-04-03",
|
| 341 |
+
"status": "full",
|
| 342 |
+
"main_language": "English",
|
| 343 |
+
"model_type": "proprietary",
|
| 344 |
+
"result_metrics": {
|
| 345 |
+
"enem_challenge": 0.8327501749475158,
|
| 346 |
+
"bluex": 0.760778859527121,
|
| 347 |
+
"oab_exams": 0.6369020501138952,
|
| 348 |
+
"assin2_sts": 0.8380176734291938,
|
| 349 |
+
"assin2_rte": 0.941176117215237,
|
| 350 |
+
"faquad_nli": 0.8360786822325283,
|
| 351 |
+
"hatebr_offensive": 0.9046145161133335,
|
| 352 |
+
"portuguese_hate_speech": 0.7406414313684444,
|
| 353 |
+
"tweetsentbr": 0.6997509880131249
|
| 354 |
+
},
|
| 355 |
+
"result_metrics_average": 0.7989678325511549,
|
| 356 |
+
"result_metrics_npm": 0.6979777100000177
|
| 357 |
+
},
|
| 358 |
+
{
|
| 359 |
+
"model": "gemini-1.5-flash-8b-001",
|
| 360 |
+
"name": "Gemini 1.5 Flash 8B (001)",
|
| 361 |
+
"link": "https://aistudio.google.com",
|
| 362 |
+
"date": "2025-04-03",
|
| 363 |
+
"status": "full",
|
| 364 |
+
"main_language": "English",
|
| 365 |
+
"model_type": "proprietary",
|
| 366 |
+
"params": 8.0,
|
| 367 |
+
"result_metrics": {
|
| 368 |
+
"enem_challenge": 0.7641707487753674,
|
| 369 |
+
"bluex": 0.6467315716272601,
|
| 370 |
+
"oab_exams": 0.5603644646924829,
|
| 371 |
+
"assin2_sts": 0.7638946799836569,
|
| 372 |
+
"assin2_rte": 0.9329452628161146,
|
| 373 |
+
"faquad_nli": 0.7937022965448601,
|
| 374 |
+
"hatebr_offensive": 0.850497640901663,
|
| 375 |
+
"portuguese_hate_speech": 0.7391317606010173,
|
| 376 |
+
"tweetsentbr": 0.7376684798923661
|
| 377 |
+
},
|
| 378 |
+
"result_metrics_average": 0.7543452117594209,
|
| 379 |
+
"result_metrics_npm": 0.6359642422837162
|
| 380 |
+
},
|
| 381 |
{
|
| 382 |
"model": "gemini-2.0-flash-001",
|
| 383 |
"name": "Gemini 2.0 Flash (001)",
|
|
|
|
| 421 |
},
|
| 422 |
"result_metrics_average": 0.8056048352614735,
|
| 423 |
"result_metrics_npm": 0.6986042497176748
|
| 424 |
+
},
|
| 425 |
+
{
|
| 426 |
+
"model": "gemini-2.5-pro-exp-03-25",
|
| 427 |
+
"name": "Gemini 2.5 Pro Experimental (0325)",
|
| 428 |
+
"link": "https://aistudio.google.com",
|
| 429 |
+
"date": "2025-04-03",
|
| 430 |
+
"status": "full",
|
| 431 |
+
"main_language": "English",
|
| 432 |
+
"model_type": "proprietary",
|
| 433 |
+
"result_metrics": {
|
| 434 |
+
"enem_challenge": 0.9769069279216235,
|
| 435 |
+
"bluex": 0.9499304589707928,
|
| 436 |
+
"oab_exams": 0.9216400911161731,
|
| 437 |
+
"assin2_sts": 0.837785744915033,
|
| 438 |
+
"assin2_rte": 0.9415510158830285,
|
| 439 |
+
"faquad_nli": 0.8738735797309651,
|
| 440 |
+
"hatebr_offensive": 0.9248478168290788,
|
| 441 |
+
"portuguese_hate_speech": 0.7336133105156697,
|
| 442 |
+
"tweetsentbr": 0.7928002469993594
|
| 443 |
+
},
|
| 444 |
+
"result_metrics_average": 0.8836610214313025,
|
| 445 |
+
"result_metrics_npm": 0.8134610556797854
|
| 446 |
+
},
|
| 447 |
+
{
|
| 448 |
+
"model": "qwen2-5-vl-72b-instruct",
|
| 449 |
+
"name": "Qwen/Qwen2.5-VL-72B-Instruct (API)",
|
| 450 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-VL-72B-Instruct",
|
| 451 |
+
"date": "2025-04-03",
|
| 452 |
+
"status": "full",
|
| 453 |
+
"main_language": "English",
|
| 454 |
+
"model_type": "chat",
|
| 455 |
+
"result_metrics": {
|
| 456 |
+
"enem_challenge": 0.8600419874037789,
|
| 457 |
+
"bluex": 0.8052851182197497,
|
| 458 |
+
"oab_exams": 0.6888382687927107,
|
| 459 |
+
"assin2_sts": 0.7595538567467497,
|
| 460 |
+
"assin2_rte": 0.9472975104201871,
|
| 461 |
+
"faquad_nli": 0.8447190882122586,
|
| 462 |
+
"hatebr_offensive": 0.8810695094657859,
|
| 463 |
+
"portuguese_hate_speech": 0.769596419318135,
|
| 464 |
+
"tweetsentbr": 0.5644757075411895
|
| 465 |
+
},
|
| 466 |
+
"result_metrics_average": 0.7912086073467273,
|
| 467 |
+
"result_metrics_npm": 0.6888261361422966
|
| 468 |
+
},
|
| 469 |
+
{
|
| 470 |
+
"model": "qwen2-5-72b-instruct",
|
| 471 |
+
"name": "Qwen/Qwen2.5-72B-Instruct (API)",
|
| 472 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-72B-Instruct",
|
| 473 |
+
"date": "2025-04-03",
|
| 474 |
+
"status": "full",
|
| 475 |
+
"main_language": "English",
|
| 476 |
+
"model_type": "chat",
|
| 477 |
+
"result_metrics": {
|
| 478 |
+
"enem_challenge": 0.8432470258922323,
|
| 479 |
+
"bluex": 0.780250347705146,
|
| 480 |
+
"oab_exams": 0.675626423690205,
|
| 481 |
+
"assin2_sts": 0.8230708844558656,
|
| 482 |
+
"assin2_rte": 0.9509720145268106,
|
| 483 |
+
"faquad_nli": 0.8194444444444444,
|
| 484 |
+
"hatebr_offensive": 0.8810033427242816,
|
| 485 |
+
"portuguese_hate_speech": 0.7601866578782712,
|
| 486 |
+
"tweetsentbr": 0.7620172222071487
|
| 487 |
+
},
|
| 488 |
+
"result_metrics_average": 0.8106464848360451,
|
| 489 |
+
"result_metrics_npm": 0.7142994872542282
|
| 490 |
+
},
|
| 491 |
+
{
|
| 492 |
+
"model": "qwen2-5-vl-32b-instruct",
|
| 493 |
+
"name": "Qwen/Qwen2.5-VL-32B-Instruct (API)",
|
| 494 |
+
"link": "https://huggingface.co/Qwen/Qwen2.5-VL-32B-Instruct",
|
| 495 |
+
"date": "2025-04-03",
|
| 496 |
+
"status": "full",
|
| 497 |
+
"main_language": "English",
|
| 498 |
+
"model_type": "chat",
|
| 499 |
+
"result_metrics": {
|
| 500 |
+
"enem_challenge": 0.8600419874037789,
|
| 501 |
+
"bluex": 0.8052851182197497,
|
| 502 |
+
"oab_exams": 0.6888382687927107,
|
| 503 |
+
"assin2_sts": 0.7780549055529008,
|
| 504 |
+
"assin2_rte": 0.9472975104201871,
|
| 505 |
+
"faquad_nli": 0.8447190882122586,
|
| 506 |
+
"hatebr_offensive": 0.8810695094657859,
|
| 507 |
+
"portuguese_hate_speech": 0.769596419318135,
|
| 508 |
+
"tweetsentbr": 0.7027408707999051
|
| 509 |
+
},
|
| 510 |
+
"result_metrics_average": 0.8086270753539346,
|
| 511 |
+
"result_metrics_npm": 0.7137431116807307
|
| 512 |
+
},
|
| 513 |
+
{
|
| 514 |
+
"model": "qwen-turbo-2024-11-01",
|
| 515 |
+
"name": "Qwen-Turbo (2024-11-01)",
|
| 516 |
+
"link": "https://www.alibabacloud.com/en/product/modelstudio",
|
| 517 |
+
"date": "2025-04-03",
|
| 518 |
+
"status": "full",
|
| 519 |
+
"main_language": "English",
|
| 520 |
+
"model_type": "proprietary",
|
| 521 |
+
"result_metrics": {
|
| 522 |
+
"enem_challenge": 0.7795661301609517,
|
| 523 |
+
"bluex": 0.7079276773296245,
|
| 524 |
+
"oab_exams": 0.6091116173120729,
|
| 525 |
+
"assin2_sts": 0.7640477700456898,
|
| 526 |
+
"assin2_rte": 0.9260451969385788,
|
| 527 |
+
"faquad_nli": 0.8128063725490196,
|
| 528 |
+
"hatebr_offensive": 0.8567933277676292,
|
| 529 |
+
"portuguese_hate_speech": 0.7239183383094245,
|
| 530 |
+
"tweetsentbr": 0.7038360447972195
|
| 531 |
+
},
|
| 532 |
+
"result_metrics_average": 0.7648947194678011,
|
| 533 |
+
"result_metrics_npm": 0.6490441260447987
|
| 534 |
}
|
| 535 |
]
|
model_list.txt
CHANGED
|
@@ -84,6 +84,7 @@
|
|
| 84 |
- EleutherAI/gpt-neo-125m
|
| 85 |
- EleutherAI/gpt-neo-2.7B
|
| 86 |
- EleutherAI/gpt-neox-20b
|
|
|
|
| 87 |
- EleutherAI/pythia-12b
|
| 88 |
- EleutherAI/pythia-14m
|
| 89 |
- EleutherAI/pythia-160m
|
|
@@ -106,8 +107,11 @@
|
|
| 106 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
| 107 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
| 108 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
|
|
|
| 109 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
| 110 |
- Eurdem/Defne-llama3.1-8B
|
|
|
|
|
|
|
| 111 |
- FuseAI/FuseChat-7B-VaRM
|
| 112 |
- FuseAI/OpenChat-3.5-7B-Solar
|
| 113 |
- GeneZC/MiniChat-1.5-3B
|
|
@@ -172,6 +176,7 @@
|
|
| 172 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
| 173 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
| 174 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
|
|
|
| 175 |
- MagusCorp/legislinho
|
| 176 |
- MarinaraSpaghetti/NemoReRemix-12B
|
| 177 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
@@ -183,6 +188,10 @@
|
|
| 183 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
| 184 |
- MaziyarPanahi/Mistral-7B-v0.3
|
| 185 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
|
|
|
|
|
|
|
|
|
|
|
|
| 186 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
| 187 |
- MulaBR/Mula-4x160-v0.1
|
| 188 |
- MulaBR/Mula-8x160-v0.1
|
|
@@ -192,6 +201,8 @@
|
|
| 192 |
- NTQAI/Nxcode-CQ-7B-orpo
|
| 193 |
- Nexusflow/Starling-LM-7B-beta
|
| 194 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
|
|
|
|
|
|
| 195 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
| 196 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
| 197 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
@@ -264,6 +275,7 @@
|
|
| 264 |
- Qwen/Qwen2.5-1.5B-Instruct
|
| 265 |
- Qwen/Qwen2.5-14B
|
| 266 |
- Qwen/Qwen2.5-14B-Instruct
|
|
|
|
| 267 |
- Qwen/Qwen2.5-32B
|
| 268 |
- Qwen/Qwen2.5-32B-Instruct
|
| 269 |
- Qwen/Qwen2.5-3B
|
|
@@ -278,8 +290,13 @@
|
|
| 278 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
| 279 |
- Qwen/Qwen2.5-Coder-7B
|
| 280 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
|
|
|
|
|
|
|
|
|
|
|
|
| 281 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
| 282 |
- Ramikan-BR/Qwen2-0.5B-v5
|
|
|
|
| 283 |
- Ramikan-BR/TiamaPY-v30
|
| 284 |
- Ramikan-BR/TiamaPY-v31
|
| 285 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
@@ -291,7 +308,9 @@
|
|
| 291 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
| 292 |
- Sakalti/Oxyge1-33B
|
| 293 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
|
|
|
| 294 |
- Sao10K/L3-8B-Niitama-v1
|
|
|
|
| 295 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
| 296 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
| 297 |
- SeaLLMs/SeaLLM-7B-v2
|
|
@@ -318,6 +337,7 @@
|
|
| 318 |
- TheBloke/wizardLM-7B-HF
|
| 319 |
- TheBloke/zephyr-7B-beta-GPTQ
|
| 320 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
|
|
|
| 321 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
| 322 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 323 |
- TinyLlama/TinyLlama_v1.1
|
|
@@ -339,6 +359,7 @@
|
|
| 339 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
| 340 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
| 341 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
|
|
|
| 342 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
| 343 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
| 344 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
@@ -368,9 +389,12 @@
|
|
| 368 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
| 369 |
- abacusai/Liberated-Qwen1.5-14B
|
| 370 |
- abacusai/Llama-3-Smaug-8B
|
|
|
|
| 371 |
- abacusai/Smaug-34B-v0.1
|
| 372 |
- abacusai/Smaug-72B-v0.1
|
| 373 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
|
|
|
|
|
|
| 374 |
- abhishek/autotrain-llama3-orpo-v2
|
| 375 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
| 376 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
@@ -416,6 +440,7 @@
|
|
| 416 |
- arcee-ai/Arcee-Spark
|
| 417 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
| 418 |
- arcee-ai/SuperNova-Medius
|
|
|
|
| 419 |
- arcee-ai/Virtuoso-Small
|
| 420 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
| 421 |
- argilla/notus-7b-v1
|
|
@@ -443,6 +468,7 @@
|
|
| 443 |
- botbot-ai/CabraLlama3-8b
|
| 444 |
- botbot-ai/CabraMistral-v3-7b-32k
|
| 445 |
- botbot-ai/CabraMixtral-8x7b
|
|
|
|
| 446 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
| 447 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
| 448 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
@@ -460,6 +486,10 @@
|
|
| 460 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
| 461 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
| 462 |
- cnmoro/Mistral-7B-Portuguese
|
|
|
|
|
|
|
|
|
|
|
|
|
| 463 |
- cognitivecomputations/WestLake-7B-v2-laser
|
| 464 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
| 465 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
@@ -481,6 +511,7 @@
|
|
| 481 |
- cognitivess/bella-1-8b
|
| 482 |
- cosmicvalor/mistral-orthogonalized
|
| 483 |
- croissantllm/CroissantLLMBase
|
|
|
|
| 484 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
| 485 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
| 486 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
@@ -544,6 +575,7 @@
|
|
| 544 |
- freewheelin/free-solar-evo-v0.1
|
| 545 |
- freewheelin/free-solar-evo-v0.11
|
| 546 |
- freewheelin/free-solar-evo-v0.13
|
|
|
|
| 547 |
- ghost-x/ghost-8b-beta
|
| 548 |
- ghost-x/ghost-8b-beta-1608
|
| 549 |
- google/gemma-1.1-2b-it
|
|
@@ -704,6 +736,7 @@
|
|
| 704 |
- microsoft/phi-1_5
|
| 705 |
- microsoft/phi-2
|
| 706 |
- microsoft/phi-4
|
|
|
|
| 707 |
- migtissera/Tess-M-v1.3
|
| 708 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
| 709 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
@@ -737,12 +770,14 @@
|
|
| 737 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
| 738 |
- mlabonne/NeuralMonarch-7B
|
| 739 |
- monilouise/opt125M_portuguese
|
|
|
|
| 740 |
- mosaicml/mpt-30b
|
| 741 |
- mosaicml/mpt-7b
|
| 742 |
- mosaicml/mpt-7b-8k
|
| 743 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
| 744 |
- natong19/Qwen2-7B-Instruct-abliterated
|
| 745 |
- nbeerbower/gemma2-gutenberg-27B
|
|
|
|
| 746 |
- nicholasKluge/Aira-2-portuguese-124M
|
| 747 |
- nicholasKluge/Aira-2-portuguese-1B7
|
| 748 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
@@ -826,6 +861,7 @@
|
|
| 826 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
| 827 |
- princeton-nlp/gemma-2-9b-it-DPO
|
| 828 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
|
|
|
| 829 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
| 830 |
- projecte-aina/FLOR-1.3B
|
| 831 |
- projecte-aina/FLOR-6.3B
|
|
@@ -904,11 +940,13 @@
|
|
| 904 |
- shadowml/BeagSake-7B
|
| 905 |
- shadowml/Mixolar-4x7b
|
| 906 |
- sometimesanotion/Lamarck-14B-v0.6
|
|
|
|
| 907 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
| 908 |
- speakleash/Bielik-11B-v2
|
| 909 |
- speakleash/Bielik-11B-v2.0-Instruct
|
| 910 |
- speakleash/Bielik-11B-v2.1-Instruct
|
| 911 |
- speakleash/Bielik-11B-v2.2-Instruct
|
|
|
|
| 912 |
- ssmits/Falcon2-5.5B-Portuguese
|
| 913 |
- ssmits/Falcon2-5.5B-multilingual
|
| 914 |
- stabilityai/stablelm-2-12b
|
|
|
|
| 84 |
- EleutherAI/gpt-neo-125m
|
| 85 |
- EleutherAI/gpt-neo-2.7B
|
| 86 |
- EleutherAI/gpt-neox-20b
|
| 87 |
+
- EleutherAI/polyglot-ko-12.8b
|
| 88 |
- EleutherAI/pythia-12b
|
| 89 |
- EleutherAI/pythia-14m
|
| 90 |
- EleutherAI/pythia-160m
|
|
|
|
| 107 |
- EpistemeAI/Fireball-Meta-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto
|
| 108 |
- EpistemeAI/Polypsyche-Llama-3.1-8B-Instruct-Agent-0.003-128K-code-ds-auto-Logic
|
| 109 |
- EpistemeAI2/Fireball-Alpaca-Llama-3.1-8B-Instruct-KTO-beta
|
| 110 |
+
- EpistemeAI2/Fireball-Alpaca-Llama3.1.06-8B-Philos
|
| 111 |
- EpistemeAI2/Fireball-Llama-3.1-8B-Philos-Relection
|
| 112 |
- Eurdem/Defne-llama3.1-8B
|
| 113 |
+
- FluxiIA/Tucaninho
|
| 114 |
+
- FluxiIA/Tucaninho_dpo
|
| 115 |
- FuseAI/FuseChat-7B-VaRM
|
| 116 |
- FuseAI/OpenChat-3.5-7B-Solar
|
| 117 |
- GeneZC/MiniChat-1.5-3B
|
|
|
|
| 176 |
- Magpie-Align/Llama-3.1-8B-Magpie-Align-SFT-v0.1
|
| 177 |
- Magpie-Align/MagpieLM-8B-Chat-v0.1
|
| 178 |
- Magpie-Align/MagpieLM-8B-SFT-v0.1
|
| 179 |
+
- MagusCorp/grpo_lora_enem_llama3_7b
|
| 180 |
- MagusCorp/legislinho
|
| 181 |
- MarinaraSpaghetti/NemoReRemix-12B
|
| 182 |
- MaziyarPanahi/Calme-4x7B-MoE-v0.1
|
|
|
|
| 188 |
- MaziyarPanahi/Mistral-7B-Instruct-v0.3
|
| 189 |
- MaziyarPanahi/Mistral-7B-v0.3
|
| 190 |
- MaziyarPanahi/Topxtral-4x7B-v0.1
|
| 191 |
+
- MaziyarPanahi/calme-2.1-qwen2-7b
|
| 192 |
+
- MaziyarPanahi/calme-2.2-qwen2-7b
|
| 193 |
+
- MaziyarPanahi/calme-2.3-qwen2-7b
|
| 194 |
+
- MaziyarPanahi/calme-2.4-qwen2-7b
|
| 195 |
- MaziyarPanahi/calme-2.7-qwen2-7b
|
| 196 |
- MulaBR/Mula-4x160-v0.1
|
| 197 |
- MulaBR/Mula-8x160-v0.1
|
|
|
|
| 201 |
- NTQAI/Nxcode-CQ-7B-orpo
|
| 202 |
- Nexusflow/Starling-LM-7B-beta
|
| 203 |
- Nos-PT/Carvalho_pt-gl-1.3B
|
| 204 |
+
- Nos-PT/Llama-Carvalho-PT
|
| 205 |
+
- Nos-PT/Llama-Carvalho-PT-GL
|
| 206 |
- NotAiLOL/Yi-1.5-dolphin-9B
|
| 207 |
- NousResearch/Hermes-2-Pro-Llama-3-8B
|
| 208 |
- NousResearch/Hermes-2-Theta-Llama-3-8B
|
|
|
|
| 275 |
- Qwen/Qwen2.5-1.5B-Instruct
|
| 276 |
- Qwen/Qwen2.5-14B
|
| 277 |
- Qwen/Qwen2.5-14B-Instruct
|
| 278 |
+
- Qwen/Qwen2.5-14B-Instruct-1M
|
| 279 |
- Qwen/Qwen2.5-32B
|
| 280 |
- Qwen/Qwen2.5-32B-Instruct
|
| 281 |
- Qwen/Qwen2.5-3B
|
|
|
|
| 290 |
- Qwen/Qwen2.5-Coder-32B-Instruct
|
| 291 |
- Qwen/Qwen2.5-Coder-7B
|
| 292 |
- Qwen/Qwen2.5-Coder-7B-Instruct
|
| 293 |
+
- Qwen/Qwen2.5-VL-32B-Instruct
|
| 294 |
+
- Qwen/Qwen2.5-VL-72B-Instruct
|
| 295 |
+
- RDson/WomboCombo-R1-Coder-14B-Preview
|
| 296 |
+
- RLHFlow/ArmoRM-Llama3-8B-v0.1
|
| 297 |
- RLHFlow/LLaMA3-iterative-DPO-final
|
| 298 |
- Ramikan-BR/Qwen2-0.5B-v5
|
| 299 |
+
- Ramikan-BR/Qwen2-0.5B-v9
|
| 300 |
- Ramikan-BR/TiamaPY-v30
|
| 301 |
- Ramikan-BR/TiamaPY-v31
|
| 302 |
- Ramikan-BR/tinyllama-coder-py-4bit-v10
|
|
|
|
| 308 |
- RubielLabarta/LogoS-7Bx2-MoE-13B-v0.2
|
| 309 |
- Sakalti/Oxyge1-33B
|
| 310 |
- SakanaAI/DiscoPOP-zephyr-7b-gemma
|
| 311 |
+
- Salesforce/SFR-Iterative-DPO-LLaMA-3-8B-R
|
| 312 |
- Sao10K/L3-8B-Niitama-v1
|
| 313 |
+
- Sao10K/L3.1-8B-Niitama-v1.1
|
| 314 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V4-32B
|
| 315 |
- Saxo/Linkbricks-Horizon-AI-Avengers-V5-32B
|
| 316 |
- SeaLLMs/SeaLLM-7B-v2
|
|
|
|
| 337 |
- TheBloke/wizardLM-7B-HF
|
| 338 |
- TheBloke/zephyr-7B-beta-GPTQ
|
| 339 |
- TheDrummer/Big-Tiger-Gemma-27B-v1
|
| 340 |
+
- TheDrummer/Gemmasutra-9B-v1
|
| 341 |
- TinyLlama/TinyLlama-1.1B-Chat-v1.0
|
| 342 |
- TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
|
| 343 |
- TinyLlama/TinyLlama_v1.1
|
|
|
|
| 359 |
- VAGOsolutions/Llama-3.1-SauerkrautLM-8b-Instruct
|
| 360 |
- VAGOsolutions/SauerkrautLM-Gemma-7b
|
| 361 |
- VAGOsolutions/SauerkrautLM-Nemo-12b-Instruct
|
| 362 |
+
- VAGOsolutions/SauerkrautLM-Phi-3-medium
|
| 363 |
- VAGOsolutions/SauerkrautLM-Qwen-32b
|
| 364 |
- ValiantLabs/Llama3.1-8B-Cobalt
|
| 365 |
- ValiantLabs/Llama3.1-8B-Fireplace2
|
|
|
|
| 389 |
- Xwin-LM/Xwin-LM-7B-V0.2
|
| 390 |
- abacusai/Liberated-Qwen1.5-14B
|
| 391 |
- abacusai/Llama-3-Smaug-8B
|
| 392 |
+
- abacusai/Slerp-CM-mist-dpo
|
| 393 |
- abacusai/Smaug-34B-v0.1
|
| 394 |
- abacusai/Smaug-72B-v0.1
|
| 395 |
- abacusai/Smaug-Llama-3-70B-Instruct
|
| 396 |
+
- abacusai/bigstral-12b-32k
|
| 397 |
+
- abacusai/bigyi-15b
|
| 398 |
- abhishek/autotrain-llama3-orpo-v2
|
| 399 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese
|
| 400 |
- adalbertojunior/Llama-3-8B-Dolphin-Portuguese-v0.2
|
|
|
|
| 440 |
- arcee-ai/Arcee-Spark
|
| 441 |
- arcee-ai/Llama-3.1-SuperNova-Lite
|
| 442 |
- arcee-ai/SuperNova-Medius
|
| 443 |
+
- arcee-ai/Virtuoso-Lite
|
| 444 |
- arcee-ai/Virtuoso-Small
|
| 445 |
- argilla/CapybaraHermes-2.5-Mistral-7B
|
| 446 |
- argilla/notus-7b-v1
|
|
|
|
| 468 |
- botbot-ai/CabraLlama3-8b
|
| 469 |
- botbot-ai/CabraMistral-v3-7b-32k
|
| 470 |
- botbot-ai/CabraMixtral-8x7b
|
| 471 |
+
- bunnycore/HyperLlama-3.1-8B
|
| 472 |
- byroneverson/Mistral-Small-Instruct-2409-abliterated
|
| 473 |
- byroneverson/Yi-1.5-9B-Chat-16K-abliterated
|
| 474 |
- byroneverson/Yi-1.5-9B-Chat-abliterated
|
|
|
|
| 486 |
- chujiezheng/Mistral7B-PairRM-SPPO-ExPO
|
| 487 |
- chujiezheng/Smaug-34B-v0.1-ExPO
|
| 488 |
- cnmoro/Mistral-7B-Portuguese
|
| 489 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-Hybrid-Reasoning
|
| 490 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v1
|
| 491 |
+
- cnmoro/Qwen2.5-0.5B-Portuguese-v2
|
| 492 |
+
- cognitivecomputations/Dolphin3.0-R1-Mistral-24B
|
| 493 |
- cognitivecomputations/WestLake-7B-v2-laser
|
| 494 |
- cognitivecomputations/WizardLM-13B-Uncensored
|
| 495 |
- cognitivecomputations/WizardLM-30B-Uncensored
|
|
|
|
| 511 |
- cognitivess/bella-1-8b
|
| 512 |
- cosmicvalor/mistral-orthogonalized
|
| 513 |
- croissantllm/CroissantLLMBase
|
| 514 |
+
- deepseek-ai/DeepSeek-R1-Distill-Llama-70B
|
| 515 |
- deepseek-ai/DeepSeek-R1-Distill-Llama-8B
|
| 516 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
|
| 517 |
- deepseek-ai/DeepSeek-R1-Distill-Qwen-14B
|
|
|
|
| 575 |
- freewheelin/free-solar-evo-v0.1
|
| 576 |
- freewheelin/free-solar-evo-v0.11
|
| 577 |
- freewheelin/free-solar-evo-v0.13
|
| 578 |
+
- gaverfraxz/Meta-Llama-3.1-8B-Instruct-HalfAbliterated-TIES
|
| 579 |
- ghost-x/ghost-8b-beta
|
| 580 |
- ghost-x/ghost-8b-beta-1608
|
| 581 |
- google/gemma-1.1-2b-it
|
|
|
|
| 736 |
- microsoft/phi-1_5
|
| 737 |
- microsoft/phi-2
|
| 738 |
- microsoft/phi-4
|
| 739 |
+
- migtissera/Tess-3-Mistral-Nemo-12B
|
| 740 |
- migtissera/Tess-M-v1.3
|
| 741 |
- migtissera/Tess-v2.5-Gemma-2-27B-alpha
|
| 742 |
- migtissera/Tess-v2.5-Phi-3-medium-128k-14B
|
|
|
|
| 770 |
- mlabonne/NeuralDaredevil-8B-abliterated
|
| 771 |
- mlabonne/NeuralMonarch-7B
|
| 772 |
- monilouise/opt125M_portuguese
|
| 773 |
+
- monsterapi/Llama-3_1-8B-Instruct-orca-ORPO
|
| 774 |
- mosaicml/mpt-30b
|
| 775 |
- mosaicml/mpt-7b
|
| 776 |
- mosaicml/mpt-7b-8k
|
| 777 |
- natong19/Mistral-Nemo-Instruct-2407-abliterated
|
| 778 |
- natong19/Qwen2-7B-Instruct-abliterated
|
| 779 |
- nbeerbower/gemma2-gutenberg-27B
|
| 780 |
+
- nbeerbower/mistral-nemo-wissenschaft-12B
|
| 781 |
- nicholasKluge/Aira-2-portuguese-124M
|
| 782 |
- nicholasKluge/Aira-2-portuguese-1B7
|
| 783 |
- nicholasKluge/Aira-2-portuguese-560M
|
|
|
|
| 861 |
- princeton-nlp/Mistral-7B-Instruct-SimPO
|
| 862 |
- princeton-nlp/gemma-2-9b-it-DPO
|
| 863 |
- princeton-nlp/gemma-2-9b-it-SimPO
|
| 864 |
+
- prithivMLmods/Megatron-Opus-14B-Exp
|
| 865 |
- prithivMLmods/Qwen2.5-14B-DeepSeek-R1-1M
|
| 866 |
- projecte-aina/FLOR-1.3B
|
| 867 |
- projecte-aina/FLOR-6.3B
|
|
|
|
| 940 |
- shadowml/BeagSake-7B
|
| 941 |
- shadowml/Mixolar-4x7b
|
| 942 |
- sometimesanotion/Lamarck-14B-v0.6
|
| 943 |
+
- sometimesanotion/Lamarck-14B-v0.7
|
| 944 |
- sometimesanotion/Qwen2.5-14B-Vimarckoso-v3
|
| 945 |
- speakleash/Bielik-11B-v2
|
| 946 |
- speakleash/Bielik-11B-v2.0-Instruct
|
| 947 |
- speakleash/Bielik-11B-v2.1-Instruct
|
| 948 |
- speakleash/Bielik-11B-v2.2-Instruct
|
| 949 |
+
- speakleash/Bielik-11B-v2.3-Instruct
|
| 950 |
- ssmits/Falcon2-5.5B-Portuguese
|
| 951 |
- ssmits/Falcon2-5.5B-multilingual
|
| 952 |
- stabilityai/stablelm-2-12b
|
update_models_in_readme.py
CHANGED
|
@@ -18,8 +18,12 @@ import json
|
|
| 18 |
|
| 19 |
snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
|
| 20 |
all_models = []
|
| 21 |
-
|
| 22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
|
| 24 |
with open(filepath, 'r') as f:
|
| 25 |
model_data = json.load(f)
|
|
|
|
| 18 |
|
| 19 |
snapshot_download(repo_id=QUEUE_REPO, local_dir=EVAL_REQUESTS_PATH, repo_type="dataset", tqdm_class=None, etag_timeout=30)
|
| 20 |
all_models = []
|
| 21 |
+
with open('external_models_results.json', 'r') as f:
|
| 22 |
+
external_models = json.load(f)
|
| 23 |
+
for model in external_models:
|
| 24 |
+
if 'huggingface.co/' in model['link']:
|
| 25 |
+
all_models.append(model['link'].split('huggingface.co/')[-1])
|
| 26 |
+
|
| 27 |
for filepath in glob.glob(os.path.join(EVAL_REQUESTS_PATH, '**/*.json'), recursive=True):
|
| 28 |
with open(filepath, 'r') as f:
|
| 29 |
model_data = json.load(f)
|