jerryzh168 commited on
Commit
d2d3073
·
verified ·
1 Parent(s): 72d7c17

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -4
README.md CHANGED
@@ -185,6 +185,7 @@ and use a token with write access, from https://huggingface.co/settings/tokens
185
 
186
  # Model Quality
187
  We rely on [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) to evaluate the quality of the quantized model. Here we only run on mmlu for sanity check.
 
188
 
189
  | Benchmark | | |
190
  |----------------------------------|----------------|---------------------------|
@@ -196,18 +197,32 @@ We rely on [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-h
196
  <details>
197
  <summary> Reproduce Model Quality Results </summary>
198
 
 
199
  Need to install lm-eval from source:
200
  https://github.com/EleutherAI/lm-evaluation-harness#install
201
 
202
- ## baseline
203
  ```Shell
 
204
  lm_eval --model hf --model_args pretrained=google/gemma-3-12b-it --tasks mmlu --device cuda:0 --batch_size 8
205
  ```
206
 
207
- ## FP8
 
 
 
208
  ```Shell
209
- export MODEL=pytorch/gemma-3-12b-it-FP8
210
- lm_eval --model hf --model_args pretrained=$MODEL --tasks mmlu --device cuda:0 --batch_size 8
 
 
 
 
 
 
 
 
 
 
211
  ```
212
  </details>
213
 
 
185
 
186
  # Model Quality
187
  We rely on [lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) to evaluate the quality of the quantized model. Here we only run on mmlu for sanity check.
188
+ We also rely on [lmms-eval](https://github.com/EvolvingLMMs-Lab/lmms-eval/) for multi-modal quality eval. We only tested on chartqa for sanity check.
189
 
190
  | Benchmark | | |
191
  |----------------------------------|----------------|---------------------------|
 
197
  <details>
198
  <summary> Reproduce Model Quality Results </summary>
199
 
200
+ ## language eval
201
  Need to install lm-eval from source:
202
  https://github.com/EleutherAI/lm-evaluation-harness#install
203
 
 
204
  ```Shell
205
+ export MODEL=google/gemma-3-12b-it # or pytorch/gemma-3-12b-it-FP8
206
  lm_eval --model hf --model_args pretrained=google/gemma-3-12b-it --tasks mmlu --device cuda:0 --batch_size 8
207
  ```
208
 
209
+ ## multi-modal eval
210
+ Need to install lmms-eval from source:
211
+ `pip install git+https://github.com/EvolvingLMMs-Lab/lmms-eval.git`
212
+
213
  ```Shell
214
+ NUM_PROCESSES=8
215
+ MAIN_PORT=12345
216
+ MODEL_ID=google/gemma-3-12b-it # or pytorch/gemma-3-12b-it-FP8
217
+ TASKS=chartqa # or tasks from https://github.com/EvolvingLMMs-Lab/lmms-eval/tree/main/lmms_eval/models/simple
218
+ BATCH_SIZE=32
219
+ OUTPUT_PATH=./logs/
220
+
221
+ accelerate launch --num_processes "${NUM_PROCESSES}" --main_process_port "${MAIN_PORT}" -m lmms_eval \
222
+ --model gemma3 \
223
+ --model_args "pretrained=${MODEL_ID}" \
224
+ --tasks "${TASKS}" \
225
+ --batch_size "${BATCH_SIZE}" --output_path "${OUTPUT_PATH}"
226
  ```
227
  </details>
228