reset the path in config files

This commit is contained in:
Deyao Zhu 2023-10-13 17:11:11 +03:00
parent b700afd18b
commit a3709f52e0
6 changed files with 6 additions and 9 deletions

View File

@ -5,8 +5,7 @@ model:
end_sym: "###" end_sym: "###"
low_resource: True low_resource: True
prompt_template: '###Human: {} ###Assistant: ' prompt_template: '###Human: {} ###Assistant: '
ckpt: '/home/zhud/weights/minigpt4/prerained_minigpt4_7b.pth' ckpt: 'please set this value to the path of pretrained checkpoint'
llama_model: "/home/zhud/weights/vicuna-7b"
datasets: datasets:

View File

@ -5,8 +5,7 @@ model:
end_sym: "</s>" end_sym: "</s>"
low_resource: True low_resource: True
prompt_template: '[INST] {} [/INST] ' prompt_template: '[INST] {} [/INST] '
ckpt: '/home/zhud/weights/minigpt4/pretrained_minigpt4_llama2_7b.pth' ckpt: 'please set this value to the path of pretrained checkpoint'
llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update"
datasets: datasets:

View File

@ -5,8 +5,7 @@ model:
end_sym: "</s>" end_sym: "</s>"
low_resource: True low_resource: True
prompt_template: '[INST] {} [/INST]' prompt_template: '[INST] {} [/INST]'
ckpt: '/home/zhud/c2090/minigpt4_ckpt/448_conversation_correct_best_v7_ablation1_v5_v6/20231007035/checkpoint_35.pth' ckpt: 'please set this value to the path of pretrained checkpoint'
llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update"
lora_r: 64 lora_r: 64
lora_alpha: 16 lora_alpha: 16

View File

@ -12,7 +12,7 @@ model:
# generation configs # generation configs
prompt: "" prompt: ""
llama_model: "/path/to/llama2/weight" llama_model: "please set this value to the path of llama2-chat-7b"
preprocess: preprocess:
vis_processor: vis_processor:

View File

@ -15,7 +15,7 @@ model:
# generation configs # generation configs
prompt: "" prompt: ""
llama_model: "/path/to/vicuna/weight" llama_model: "please set this value to the path of vicuna model"
preprocess: preprocess:
vis_processor: vis_processor:

View File

@ -11,7 +11,7 @@ model:
# generation configs # generation configs
prompt: "" prompt: ""
llama_model: "/path/to/llama2/weight" llama_model: "please set this value to the path of llama2-chat-7b"
lora_r: 64 lora_r: 64
lora_alpha: 16 lora_alpha: 16