From ab520c89fc76d14b623aeea0a8e3bd5134ed36fd Mon Sep 17 00:00:00 2001 From: Xiang Li <44761952+lx709@users.noreply.github.com> Date: Mon, 23 Oct 2023 21:41:39 +0300 Subject: [PATCH 1/3] Update README_MINIGPTv2_FINETUNE.md --- dataset/README_MINIGPTv2_FINETUNE.md | 10 +++------- 1 file changed, 3 insertions(+), 7 deletions(-) diff --git a/dataset/README_MINIGPTv2_FINETUNE.md b/dataset/README_MINIGPTv2_FINETUNE.md index 3d09ffe..f658adc 100644 --- a/dataset/README_MINIGPTv2_FINETUNE.md +++ b/dataset/README_MINIGPTv2_FINETUNE.md @@ -26,7 +26,7 @@ After downloading all of them, organize the data as follows in `./playground/dat ### TextCaps ### RefCOCO, RefCOCO+, RefCOCOg -Makesure you have the COCO 2014 images first. +Make sure you have the COCO 2014 images first. Then, download RefCOCO, RefCOCO+, and RefCOCOg annotation files in the following links. @@ -71,14 +71,10 @@ Similarly, set **ann_path** in all the following configs to the above folder (Lo ### LLaVA ### TextVQA -- [train_val_images](https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip) -### GQA -- [images](https://downloads.cs.stanford.edu/nlp/data/gqa/images.zip) -- [Annotations](https://storage.googleapis.com/sfr-vision-language-research/LAVIS/datasets/gqa/testdev_balanced_questions.json) - - +Images, and question-answer pairs will be loaded during evaluation. ### GQA +Images, and question-answer pairs will be loaded during evaluation. ### OKVQA From 50df66e81e437091b7a55e68c793dee2b3b4f5f0 Mon Sep 17 00:00:00 2001 From: Xiang Li <44761952+lx709@users.noreply.github.com> Date: Mon, 23 Oct 2023 21:42:58 +0300 Subject: [PATCH 2/3] Create Evaluation.md --- dataset/Evaluation.md | 1 + 1 file changed, 1 insertion(+) create mode 100644 dataset/Evaluation.md diff --git a/dataset/Evaluation.md b/dataset/Evaluation.md new file mode 100644 index 0000000..8b13789 --- /dev/null +++ b/dataset/Evaluation.md @@ -0,0 +1 @@ + From fa19bc09f21c220c564148c9d7ed86d3c1ced523 Mon Sep 17 00:00:00 2001 From: Xiang Li <44761952+lx709@users.noreply.github.com> Date: Mon, 23 Oct 2023 21:45:22 +0300 Subject: [PATCH 3/3] Update README_MINIGPTv2_FINETUNE.md --- dataset/README_MINIGPTv2_FINETUNE.md | 22 ---------------------- 1 file changed, 22 deletions(-) diff --git a/dataset/README_MINIGPTv2_FINETUNE.md b/dataset/README_MINIGPTv2_FINETUNE.md index f658adc..280622e 100644 --- a/dataset/README_MINIGPTv2_FINETUNE.md +++ b/dataset/README_MINIGPTv2_FINETUNE.md @@ -1,26 +1,10 @@ ## Download the COCO captions, RefCOCO, RefCOCO+. RefCOCOg, visual genome, textcaps, LLaVA, gqa, AOK-VQA, OK-VQA, OCR-VQA, filtered Flickr-30k, multi-task conversation, and Unnatural instruction datasets -After downloading all of them, organize the data as follows in `./playground/data`, - -``` -├── coco -│ └── train2017 -├── gqa -│ └── images -├── ocr_vqa -│ └── images -├── textvqa -│ └── train_images -└── vg - ├── VG_100K - └── VG_100K_2 -``` ### COCO captions - [train2017](http://images.cocodataset.org/zips/train2017.zip) - ### Visual genome - [part1](https://cs.stanford.edu/people/rak248/VG_100K_2/images.zip), [part2](https://cs.stanford.edu/people/rak248/VG_100K_2/images2.zip) ### TextCaps @@ -70,12 +54,6 @@ Similarly, set **ann_path** in all the following configs to the above folder (Lo ### LLaVA -### TextVQA -Images, and question-answer pairs will be loaded during evaluation. - -### GQA -Images, and question-answer pairs will be loaded during evaluation. - ### OKVQA ### AOK-VQA