diff --git a/dataset/README_MINIGPTv2_FINETUNE.md b/dataset/README_MINIGPTv2_FINETUNE.md index f658adc..280622e 100644 --- a/dataset/README_MINIGPTv2_FINETUNE.md +++ b/dataset/README_MINIGPTv2_FINETUNE.md @@ -1,26 +1,10 @@ ## Download the COCO captions, RefCOCO, RefCOCO+. RefCOCOg, visual genome, textcaps, LLaVA, gqa, AOK-VQA, OK-VQA, OCR-VQA, filtered Flickr-30k, multi-task conversation, and Unnatural instruction datasets -After downloading all of them, organize the data as follows in `./playground/data`, - -``` -├── coco -│ └── train2017 -├── gqa -│ └── images -├── ocr_vqa -│ └── images -├── textvqa -│ └── train_images -└── vg - ├── VG_100K - └── VG_100K_2 -``` ### COCO captions - [train2017](http://images.cocodataset.org/zips/train2017.zip) - ### Visual genome - [part1](https://cs.stanford.edu/people/rak248/VG_100K_2/images.zip), [part2](https://cs.stanford.edu/people/rak248/VG_100K_2/images2.zip) ### TextCaps @@ -70,12 +54,6 @@ Similarly, set **ann_path** in all the following configs to the above folder (Lo ### LLaVA -### TextVQA -Images, and question-answer pairs will be loaded during evaluation. - -### GQA -Images, and question-answer pairs will be loaded during evaluation. - ### OKVQA ### AOK-VQA