From 023fdf11f7a5b8a8eb147eb220b8fd3ed89e96ea Mon Sep 17 00:00:00 2001 From: ThuanNaN Date: Mon, 20 Jan 2025 07:24:05 +0700 Subject: [PATCH] add finetune cfg --- train_configs/minigptv2_finetune_textvqa.yaml | 56 +++++++++++++++++++ 1 file changed, 56 insertions(+) create mode 100644 train_configs/minigptv2_finetune_textvqa.yaml diff --git a/train_configs/minigptv2_finetune_textvqa.yaml b/train_configs/minigptv2_finetune_textvqa.yaml new file mode 100644 index 0000000..9755295 --- /dev/null +++ b/train_configs/minigptv2_finetune_textvqa.yaml @@ -0,0 +1,56 @@ +model: + arch: minigpt_v2 + model_type: pretrain + max_txt_len: 1024 + image_size: 448 + end_sym: "" + llama_model: "meta-llama/Llama-2-7b-chat-hf" + ckpt: "./ckpt/checkpoint_stage3.pth" + use_grad_checkpoint: True + low_resource: True + chat_template: True + lora_r: 64 + lora_alpha: 16 + +datasets: + textvqa: + batch_size: 2 + vis_processor: + train: + name: "blip2_image_train" + image_size: 448 + text_processor: + train: + name: "blip_caption" + sample_ratio: 100 + +run: + task: image_text_pretrain + # optimizer + lr_sched: "linear_warmup_cosine_lr" + init_lr: 1e-5 + min_lr: 1e-6 + warmup_lr: 1e-6 + + weight_decay: 0.05 + max_epoch: 5 + num_workers: 6 + warmup_steps: 1000 + iters_per_epoch: 2000 + + seed: 42 + output_dir: "textvqa_outputs" + + amp: True + resume_ckpt_path: null + + evaluate: False + train_splits: ["train"] + + device: "cuda" + world_size: 1 + dist_url: "env://" + distributed: True + + wandb_log: False + job_name: minigptv2_finetune