diff --git a/eval_configs/minigpt4_eval.yaml b/eval_configs/minigpt4_eval.yaml index 1828cef..d73ba3a 100644 --- a/eval_configs/minigpt4_eval.yaml +++ b/eval_configs/minigpt4_eval.yaml @@ -5,8 +5,7 @@ model: end_sym: "###" low_resource: True prompt_template: '###Human: {} ###Assistant: ' - ckpt: '/home/zhud/weights/minigpt4/prerained_minigpt4_7b.pth' - llama_model: "/home/zhud/weights/vicuna-7b" + ckpt: 'please set this value to the path of pretrained checkpoint' datasets: diff --git a/eval_configs/minigpt4_llama2_eval.yaml b/eval_configs/minigpt4_llama2_eval.yaml index f743497..93efab1 100644 --- a/eval_configs/minigpt4_llama2_eval.yaml +++ b/eval_configs/minigpt4_llama2_eval.yaml @@ -5,8 +5,7 @@ model: end_sym: "" low_resource: True prompt_template: '[INST] {} [/INST] ' - ckpt: '/home/zhud/weights/minigpt4/pretrained_minigpt4_llama2_7b.pth' - llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update" + ckpt: 'please set this value to the path of pretrained checkpoint' datasets: diff --git a/eval_configs/minigptv2_eval.yaml b/eval_configs/minigptv2_eval.yaml index 0980178..0479f2a 100644 --- a/eval_configs/minigptv2_eval.yaml +++ b/eval_configs/minigptv2_eval.yaml @@ -5,8 +5,7 @@ model: end_sym: "" low_resource: True prompt_template: '[INST] {} [/INST]' - ckpt: '/home/zhud/c2090/minigpt4_ckpt/448_conversation_correct_best_v7_ablation1_v5_v6/20231007035/checkpoint_35.pth' - llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update" + ckpt: 'please set this value to the path of pretrained checkpoint' lora_r: 64 lora_alpha: 16 diff --git a/minigpt4/configs/models/minigpt4_llama2.yaml b/minigpt4/configs/models/minigpt4_llama2.yaml index af73f1d..fdd25e0 100644 --- a/minigpt4/configs/models/minigpt4_llama2.yaml +++ b/minigpt4/configs/models/minigpt4_llama2.yaml @@ -12,7 +12,7 @@ model: # generation configs prompt: "" - llama_model: "/path/to/llama2/weight" + llama_model: "please set this value to the path of llama2-chat-7b" preprocess: vis_processor: diff --git a/minigpt4/configs/models/minigpt4_vicuna0.yaml b/minigpt4/configs/models/minigpt4_vicuna0.yaml index 783ec4e..718054c 100644 --- a/minigpt4/configs/models/minigpt4_vicuna0.yaml +++ b/minigpt4/configs/models/minigpt4_vicuna0.yaml @@ -15,7 +15,7 @@ model: # generation configs prompt: "" - llama_model: "/path/to/vicuna/weight" + llama_model: "please set this value to the path of vicuna model" preprocess: vis_processor: diff --git a/minigpt4/configs/models/minigpt_v2.yaml b/minigpt4/configs/models/minigpt_v2.yaml index 018a18a..1d85d20 100755 --- a/minigpt4/configs/models/minigpt_v2.yaml +++ b/minigpt4/configs/models/minigpt_v2.yaml @@ -11,7 +11,7 @@ model: # generation configs prompt: "" - llama_model: "/path/to/llama2/weight" + llama_model: "please set this value to the path of llama2-chat-7b" lora_r: 64 lora_alpha: 16