Spaces:
Build error
Build error
ready for eval
Browse files
data/Qwen2-72B-Instruct_metrics.csv
CHANGED
|
@@ -1,2 +1,8 @@
|
|
| 1 |
epoch,model,accuracy,precision,recall,f1
|
| 2 |
-
0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
epoch,model,accuracy,precision,recall,f1
|
| 2 |
+
0.0,Qwen/Qwen2-72B-Instruct_torch.bfloat16_4bit_lf,0.7473333333333333,0.804122252986722,0.7473333333333333,0.7607828719113865
|
| 3 |
+
0.2,Qwen/Qwen2-72B-Instruct/checkpoint-35_torch.bfloat16_4bit_lf,0.7583333333333333,0.8199928526815756,0.7583333333333333,0.782751089787442
|
| 4 |
+
0.4,Qwen/Qwen2-72B-Instruct/checkpoint-70_torch.bfloat16_4bit_lf,0.7366666666666667,0.8224865755517643,0.7366666666666667,0.7700627366337021
|
| 5 |
+
0.6,Qwen/Qwen2-72B-Instruct/checkpoint-105_torch.bfloat16_4bit_lf,0.757,0.8253824826209251,0.757,0.784000409833628
|
| 6 |
+
0.8,Qwen/Qwen2-72B-Instruct/checkpoint-140_torch.bfloat16_4bit_lf,0.7893333333333333,0.8229104753645825,0.7893333333333333,0.8033124955993173
|
| 7 |
+
1.0,Qwen/Qwen2-72B-Instruct/checkpoint-175_torch.bfloat16_4bit_lf,0.7376666666666667,0.8243654864769323,0.7376666666666667,0.7699617360961548
|
| 8 |
+
1.2,Qwen/Qwen2-72B-Instruct/checkpoint-210_torch.bfloat16_4bit_lf,0.763,0.8318882808702871,0.763,0.7901075708186186
|
data/Qwen2-72B-Instruct_results.csv
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
notebooks/00_Data Analysis.ipynb
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
notebooks/02b_Qwen2-72B-Instruct_analysis.ipynb
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
scripts/eval-mgtv-qwen2_72b.sh
CHANGED
|
@@ -15,16 +15,47 @@ grep MemTotal /proc/meminfo
|
|
| 15 |
|
| 16 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
| 17 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
| 18 |
-
export START_EPOCH=0
|
| 19 |
export USING_LLAMA_FACTORY=true
|
|
|
|
| 20 |
export LOAD_IN_4BIT=true
|
| 21 |
-
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
| 22 |
|
|
|
|
| 23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
|
|
|
| 24 |
|
| 25 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
| 26 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
|
| 29 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 30 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
|
|
|
| 15 |
|
| 16 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
| 17 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
|
|
|
| 18 |
export USING_LLAMA_FACTORY=true
|
| 19 |
+
export USING_P1_PROMPT_TEMPLATE=false
|
| 20 |
export LOAD_IN_4BIT=true
|
|
|
|
| 21 |
|
| 22 |
+
export START_EPOCH=7
|
| 23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
| 24 |
+
export MODEL_NAME=Qwen/$Qwen2-72B-Instruct
|
| 25 |
|
| 26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
| 27 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
| 28 |
+
|
| 29 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 30 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
| 31 |
+
|
| 32 |
+
export START_EPOCH=0
|
| 33 |
+
export MODEL_PREFIX=Llama3.1-70B-Chinese-Chat
|
| 34 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
| 35 |
+
|
| 36 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
| 37 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
| 38 |
+
|
| 39 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 40 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
export LOAD_IN_4BIT=false
|
| 44 |
+
|
| 45 |
+
export MODEL_PREFIX=Llama3.1-8B-Chinese-Chat
|
| 46 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
| 47 |
+
|
| 48 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
| 49 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
| 50 |
+
|
| 51 |
+
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 52 |
+
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
| 53 |
+
|
| 54 |
+
export MODEL_PREFIX=Mistral-7B-v0.3-Chinese-Chat
|
| 55 |
+
export MODEL_NAME=shenzhi-wang/$MODEL_PREFIX
|
| 56 |
+
|
| 57 |
+
export LOGICAL_REASONING_RESULTS_PATH=data/${MODEL_PREFIX}_results.csv
|
| 58 |
+
export ADAPTER_PATH_BASE=llama-factory/saves/$MODEL_PREFIX
|
| 59 |
|
| 60 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 61 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|