Spaces:
Build error
Build error
llama3 p2 epoch 1 results
Browse files
llama-factory/config/llama3-8b_lora_sft_bf16-p1.yaml
CHANGED
|
@@ -25,7 +25,7 @@ output_dir: saves/llama3-8b/lora/sft_bf16_p1_full
|
|
| 25 |
logging_steps: 10
|
| 26 |
save_steps: 175
|
| 27 |
plot_loss: true
|
| 28 |
-
|
| 29 |
|
| 30 |
### train
|
| 31 |
per_device_train_batch_size: 16
|
|
|
|
| 25 |
logging_steps: 10
|
| 26 |
save_steps: 175
|
| 27 |
plot_loss: true
|
| 28 |
+
overwrite_output_dir: true
|
| 29 |
|
| 30 |
### train
|
| 31 |
per_device_train_batch_size: 16
|
results/llama3-8b_lora_sft_bf16-p2.csv
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
scripts/eval-mgtv-llama3_8b.sh
CHANGED
|
@@ -17,7 +17,6 @@ grep MemTotal /proc/meminfo
|
|
| 17 |
#cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
|
| 18 |
|
| 19 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
| 20 |
-
export START_EPOCH=1
|
| 21 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
| 22 |
|
| 23 |
#export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
|
|
@@ -27,11 +26,13 @@ export MODEL_PREFIX=llama3-8b_lora_sft_bf16
|
|
| 27 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
|
| 28 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
|
| 29 |
export USING_P1_PROMPT_TEMPLATE=true
|
|
|
|
| 30 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 31 |
-
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
| 32 |
|
| 33 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
|
| 34 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
|
| 35 |
export USING_P1_PROMPT_TEMPLATE=false
|
|
|
|
| 36 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 37 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
|
|
|
| 17 |
#cd ../LLaMA-Factory && pip install -e .[torch,bitsandbytes]
|
| 18 |
|
| 19 |
export LOGICAL_REASONING_DATA_PATH=datasets/mgtv
|
|
|
|
| 20 |
export RESIZE_TOKEN_EMBEDDINGS=true
|
| 21 |
|
| 22 |
#export MODEL_NAME=FlagAlpha/Llama3-Chinese-8B-Instruct
|
|
|
|
| 26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p1.csv
|
| 27 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p1_full
|
| 28 |
export USING_P1_PROMPT_TEMPLATE=true
|
| 29 |
+
export START_EPOCH=3
|
| 30 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 31 |
+
#python llm_toolkit/eval_logical_reasoning_all_epochs.py
|
| 32 |
|
| 33 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX-p2.csv
|
| 34 |
export ADAPTER_PATH_BASE=llama-factory/saves/llama3-8b/lora/sft_bf16_p2_full
|
| 35 |
export USING_P1_PROMPT_TEMPLATE=false
|
| 36 |
+
export START_EPOCH=1
|
| 37 |
echo "Eval $MODEL_NAME with $ADAPTER_PATH_BASE"
|
| 38 |
python llm_toolkit/eval_logical_reasoning_all_epochs.py
|