-
Notifications
You must be signed in to change notification settings - Fork 1
Expand file tree
/
Copy pathrun_eval.sh
More file actions
executable file
·35 lines (32 loc) · 1009 Bytes
/
run_eval.sh
File metadata and controls
executable file
·35 lines (32 loc) · 1009 Bytes
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
#!/usr/bin/env bash
DATASET="$1"
MODEL="$2"
CHECKPOINT="$3"
DATASET_SPLIT="${4:-null}"
BATCH_SIZE="${5:-8}"
GPU_MEMORY_UTILIZATION="${6:-0.5}"
SEED="${7:-null}"
echo "Running evaluation script with the following arguments:
DATASET=$DATASET
MODEL=$MODEL
CHECKPOINT=$CHECKPOINT
DATASET_SPLIT=$DATASET_SPLIT
BATCH_SIZE=$BATCH_SIZE
GPU_MEMORY_UTILIZATION=$GPU_MEMORY_UTILIZATION
CUDA_VISIBLE_DEVICES=$CUDA_VISIBLE_DEVICES
SEED=$SEED
"
for max_thinking_tokens in 8192; do
python3 generate_vllm_mcqa.py model=$MODEL \
checkpoint_name=$CHECKPOINT \
cuda_devices="'$CUDA_VISIBLE_DEVICES'" \
dataset.name=$DATASET \
dataset.split=$DATASET_SPLIT \
model.llm="Qwen/Qwen3-4B-Thinking-2507" \
model.llm_embedding_dim=2560 \
batch_size=$BATCH_SIZE \
seed=$SEED \
model.gpu_memory_utilization=$GPU_MEMORY_UTILIZATION \
model.max_tokens=16384 \
model.max_thinking_tokens=$max_thinking_tokens
done