Commit 
							
							·
						
						5c15b90
	
1
								Parent(s):
							
							8af01af
								
Adding Evaluation Results
Browse filesThis is an automated PR created with https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr
The purpose of this PR is to add evaluation results from the Open LLM Leaderboard to your model card.
If you encounter any issues, please report them to https://huggingface.co/spaces/Weyaxi/open-llm-leaderboard-results-pr/discussions
    	
        README.md
    CHANGED
    
    | @@ -232,4 +232,17 @@ If you find Zephyr-7B-β is useful in your work, please cite it with: | |
| 232 | 
             
                  archivePrefix={arXiv},
         | 
| 233 | 
             
                  primaryClass={cs.LG}
         | 
| 234 | 
             
            }
         | 
| 235 | 
            -
            ```
         | 
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 232 | 
             
                  archivePrefix={arXiv},
         | 
| 233 | 
             
                  primaryClass={cs.LG}
         | 
| 234 | 
             
            }
         | 
| 235 | 
            +
            ```
         | 
| 236 | 
            +
            # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
         | 
| 237 | 
            +
            Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_HuggingFaceH4__zephyr-7b-beta)
         | 
| 238 | 
            +
             | 
| 239 | 
            +
            | Metric                | Value                     |
         | 
| 240 | 
            +
            |-----------------------|---------------------------|
         | 
| 241 | 
            +
            | Avg.                  | 52.15   |
         | 
| 242 | 
            +
            | ARC (25-shot)         | 62.03          |
         | 
| 243 | 
            +
            | HellaSwag (10-shot)   | 84.36    |
         | 
| 244 | 
            +
            | MMLU (5-shot)         | 61.07         |
         | 
| 245 | 
            +
            | TruthfulQA (0-shot)   | 57.45   |
         | 
| 246 | 
            +
            | Winogrande (5-shot)   | 77.74   |
         | 
| 247 | 
            +
            | GSM8K (5-shot)        | 12.74        |
         | 
| 248 | 
            +
            | DROP (3-shot)         | 9.66         |
         | 

