Spaces:
Build error
Build error
change BATCH_SIZE to 1 for qwen2-72b eval
Browse files
llm_toolkit/eval_logical_reasoning.py
CHANGED
|
@@ -28,6 +28,7 @@ test_data = os.getenv("TEST_DATA", None)
|
|
| 28 |
using_llama_factory = os.getenv("USING_LLAMA_FACTORY") == "true"
|
| 29 |
max_new_tokens = int(os.getenv("MAX_NEW_TOKENS", 16))
|
| 30 |
repetition_penalty = float(os.getenv("REPETITION_PENALTY", 1.0))
|
|
|
|
| 31 |
|
| 32 |
dtype = (
|
| 33 |
torch.float32
|
|
@@ -84,7 +85,7 @@ predictions = eval_model(
|
|
| 84 |
datasets["test"],
|
| 85 |
max_new_tokens=max_new_tokens,
|
| 86 |
repetition_penalty=repetition_penalty,
|
| 87 |
-
batch_size=
|
| 88 |
)
|
| 89 |
|
| 90 |
gpu_stats = torch.cuda.get_device_properties(0)
|
|
|
|
| 28 |
using_llama_factory = os.getenv("USING_LLAMA_FACTORY") == "true"
|
| 29 |
max_new_tokens = int(os.getenv("MAX_NEW_TOKENS", 16))
|
| 30 |
repetition_penalty = float(os.getenv("REPETITION_PENALTY", 1.0))
|
| 31 |
+
batch_size = int(os.getenv("BATCH_SIZE", 2))
|
| 32 |
|
| 33 |
dtype = (
|
| 34 |
torch.float32
|
|
|
|
| 85 |
datasets["test"],
|
| 86 |
max_new_tokens=max_new_tokens,
|
| 87 |
repetition_penalty=repetition_penalty,
|
| 88 |
+
batch_size=batch_size,
|
| 89 |
)
|
| 90 |
|
| 91 |
gpu_stats = torch.cuda.get_device_properties(0)
|
notebooks/00_Data Analysis.ipynb
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
scripts/eval-mgtv-qwen2_72b.sh
CHANGED
|
@@ -19,9 +19,10 @@ export USING_LLAMA_FACTORY=true
|
|
| 19 |
export USING_P1_PROMPT_TEMPLATE=false
|
| 20 |
export LOAD_IN_4BIT=true
|
| 21 |
|
| 22 |
-
export START_EPOCH=
|
| 23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
| 24 |
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
|
|
|
| 25 |
|
| 26 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
| 27 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|
|
|
|
| 19 |
export USING_P1_PROMPT_TEMPLATE=false
|
| 20 |
export LOAD_IN_4BIT=true
|
| 21 |
|
| 22 |
+
export START_EPOCH=7
|
| 23 |
export MODEL_PREFIX=Qwen2-72B-Instruct_p2
|
| 24 |
export MODEL_NAME=Qwen/Qwen2-72B-Instruct
|
| 25 |
+
export BATCH_SIZE=1
|
| 26 |
|
| 27 |
export LOGICAL_REASONING_RESULTS_PATH=results/$MODEL_PREFIX.csv
|
| 28 |
export ADAPTER_PATH_BASE=llama-factory/saves/Qwen2-72B-Instruct
|