Skip to content

Latest commit

 

History

History
35 lines (21 loc) · 1.14 KB

Evaluation.md

File metadata and controls

35 lines (21 loc) · 1.14 KB

Evaluation

VLMEvalKit

We use VLMEVALKIT as the evaluation tools. Please refer to QuickStart for installation.

We evaluate the models with scripts in evaluation.sh. You could modify the parameters for evaluating different benchmarks.

You should use the file run.py to replace with the original run file to evaluate the model.

eval_dataset="MMVet" # need openai api key
eval_dataset="MME MMBench_DEV_EN"

# set the llava-path to the actual path of your convllava checkpoint

We would contribute the VLMEVALKIT to support our model soon.

lmms-eval

If you want to use lmms-eval to evaluate the model. You need to first install the package:

git clone https://github.com/EvolvingLMMs-Lab/lmms-eval
cd lmms-eval
pip install -e .

You should use the file eval-lmms.sh to evaluate the model. You could modify the parameters for evaluating different benchmarks.

RefCOCO

If you are interested in RefCOCO, we provide the code in refcoco.sh.