MiniGPT-4/eval_scripts/EVAL_README.md

105 lines
3.3 KiB
Markdown
Raw Normal View History

2023-10-31 01:56:49 +00:00
## Evaluation Instruction for MiniGPT-v2
### Data preparation
2023-10-31 02:37:37 +00:00
Images download
Image source | Download path
--- | :---:
OKVQA| <a href="https://drive.google.com/drive/folders/1jxIgAhtaLu_YqnZEl8Ym11f7LhX3nptN?usp=sharing">annotations</a> &nbsp;&nbsp; <a href="http://images.cocodataset.org/zips/train2017.zip"> images</a>
gqa | <a href="https://drive.google.com/drive/folders/1-dF-cgFwstutS4qq2D9CFQTDS0UTmIft?usp=drive_link">annotations</a> &nbsp;&nbsp; <a href="https://downloads.cs.stanford.edu/nlp/data/gqa/images.zip">images</a>
hateful meme | <a href="https://github.com/faizanahemad/facebook-hateful-memes">images and annotations</a>
iconqa | <a href="https://iconqa.github.io/#download">images and annotation</a>
vizwiz | <a href="https://vizwiz.org/tasks-and-datasets/vqa/">images and annotation</a>
RefCOCO | <a href="https://bvisionweb1.cs.unc.edu/licheng/referit/data/refcoco.zip"> annotations </a>
RefCOCO+ | <a href="https://bvisionweb1.cs.unc.edu/licheng/referit/data/refcoco+.zip"> annotations </a>
RefCOCOg | <a href="https://bvisionweb1.cs.unc.edu/licheng/referit/data/refcocog.zip"> annotations </a>
### Evaluation dataset structure
2023-10-31 01:56:49 +00:00
2023-10-31 05:03:07 +00:00
```
${MINIGPTv2_EVALUATION_DATASET}
├── gqa
│ └── test_balanced_questions.json
│ ├── testdev_balanced_questions.json
│ ├── gqa_images
├── hateful_meme
│ └── hm_images
│ ├── dev.jsonl
├── iconvqa
│ └── iconvqa_images
│ ├── choose_text_val.json
├── vizwiz
│ └── vizwiz_images
│ ├── val.json
├── vsr
│ └── vsr_images
├── okvqa
│ ├── okvqa_test_split.json
│ ├── mscoco_val2014_annotations_clean.json
│ ├── OpenEnded_mscoco_val2014_questions_clean.json
├── refcoco
│ └── instances.json
│ ├── refs(google).p
│ ├── refs(unc).p
├── refcoco+
│ └── instances.json
│ ├── refs(unc).p
├── refercocog
│ └── instances.json
│ ├── refs(google).p
│ ├── refs(und).p
...
```
2023-10-31 01:56:49 +00:00
### environment setup
```
export PYTHONPATH=$PYTHONPATH:/path/to/directory/of/MiniGPT-4
```
2023-11-01 06:40:27 +00:00
### config file setup
2023-11-01 06:33:48 +00:00
Set **llama_model** to the path of LLaMA model.
Set **ckpt** to the path of our pretrained model.
Set **eval_file_path** to the path of the annotation files for the evaluation data.
Set **img_path** to the path of the images.
Set **save_path** to the path of saving evaluation output.
2023-11-01 06:40:27 +00:00
in [minigpt4/eval_configs/minigptv2_benchmark_evaluation.yaml](../minigpt4/eval_configs/minigptv2_benchmark_evaluation.yaml)
2023-11-01 06:33:48 +00:00
2023-10-31 01:56:49 +00:00
### start evalauting RefCOCO, RefCOCO+, RefCOCOg
2023-10-31 06:13:47 +00:00
port=port_number
2023-11-01 06:33:48 +00:00
cfg_path=/path/to/eval_configs/minigptv2_benchmark_evaluation.yaml
2023-10-31 01:56:49 +00:00
2023-11-01 06:40:27 +00:00
dataset names:
| refcoco | refcoco+ | refcocog |
| ------- | -------- | -------- |
2023-10-31 01:56:49 +00:00
```
torchrun --master-port ${port} --nproc_per_node 1 eval_ref.py \
2023-11-01 08:05:51 +00:00
--cfg-path ${cfg_path} --dataset dataset_name --resample
2023-10-31 01:56:49 +00:00
```
### start evaluating visual question answering
2023-10-31 06:13:47 +00:00
port=port_number
2023-11-01 06:33:48 +00:00
cfg_path=/path/to/eval_configs/minigptv2_benchmark_evaluation.yaml
2023-10-31 01:56:49 +00:00
2023-11-01 06:41:29 +00:00
dataset names:
| okvqa | vizwiz | iconvqa | gqa | vsr | hm |
| ------- | -------- | -------- |-------- | -------- | -------- |
2023-10-31 01:56:49 +00:00
```
2023-10-31 06:04:43 +00:00
torchrun --master-port ${port} --nproc_per_node 1 eval_vqa.py \
2023-11-01 06:40:27 +00:00
--cfg-path ${cfg_path} --dataset dataset_name
2023-10-31 01:56:49 +00:00
```