|
|
||
|---|---|---|
| .. | ||
| README.md | ||
| bench_hf.py | ||
| bench_sglang.py | ||
| data_utils.py | ||
| eval_utils.py | ||
| prompt_format.yaml | ||
README.md
Run evaluation
Evaluate sglang
Host the VLM:
python -m sglang.launch_server --model-path Qwen/Qwen2-VL-7B-Instruct --chat-template qwen2-vl --port 30000
Benchmark:
python benchmark/mmmu/bench_sglang.py --port 30000
It's recommended to reduce the memory usage by appending something ike --mem-fraction-static 0.6 to the command above.
Evaluate hf
python benchmark/mmmu/bench_hf.py --model-path Qwen/Qwen2-VL-7B-Instruct
Some popular model results:
- Qwen/Qwen2-VL-2B-Instruct: 0.241
- Qwen/Qwen2-VL-7B-Instruct: 0.255
- Qwen/Qwen2.5-VL-3B-Instruct: 0.245
- Qwen/Qwen2.5-VL-7B-Instruct: 0.242