mirror of
https://github.com/Vision-CAIR/MiniGPT-4.git
synced 2025-04-05 10:30:45 +00:00
reset the path in config files
This commit is contained in:
parent
b700afd18b
commit
a3709f52e0
@ -5,8 +5,7 @@ model:
|
|||||||
end_sym: "###"
|
end_sym: "###"
|
||||||
low_resource: True
|
low_resource: True
|
||||||
prompt_template: '###Human: {} ###Assistant: '
|
prompt_template: '###Human: {} ###Assistant: '
|
||||||
ckpt: '/home/zhud/weights/minigpt4/prerained_minigpt4_7b.pth'
|
ckpt: 'please set this value to the path of pretrained checkpoint'
|
||||||
llama_model: "/home/zhud/weights/vicuna-7b"
|
|
||||||
|
|
||||||
|
|
||||||
datasets:
|
datasets:
|
||||||
|
@ -5,8 +5,7 @@ model:
|
|||||||
end_sym: "</s>"
|
end_sym: "</s>"
|
||||||
low_resource: True
|
low_resource: True
|
||||||
prompt_template: '[INST] {} [/INST] '
|
prompt_template: '[INST] {} [/INST] '
|
||||||
ckpt: '/home/zhud/weights/minigpt4/pretrained_minigpt4_llama2_7b.pth'
|
ckpt: 'please set this value to the path of pretrained checkpoint'
|
||||||
llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update"
|
|
||||||
|
|
||||||
|
|
||||||
datasets:
|
datasets:
|
||||||
|
@ -5,8 +5,7 @@ model:
|
|||||||
end_sym: "</s>"
|
end_sym: "</s>"
|
||||||
low_resource: True
|
low_resource: True
|
||||||
prompt_template: '[INST] {} [/INST]'
|
prompt_template: '[INST] {} [/INST]'
|
||||||
ckpt: '/home/zhud/c2090/minigpt4_ckpt/448_conversation_correct_best_v7_ablation1_v5_v6/20231007035/checkpoint_35.pth'
|
ckpt: 'please set this value to the path of pretrained checkpoint'
|
||||||
llama_model: "/ibex/project/c2133/llama_v2/llama-2-7b-chat-pytorch_update"
|
|
||||||
lora_r: 64
|
lora_r: 64
|
||||||
lora_alpha: 16
|
lora_alpha: 16
|
||||||
|
|
||||||
|
@ -12,7 +12,7 @@ model:
|
|||||||
# generation configs
|
# generation configs
|
||||||
prompt: ""
|
prompt: ""
|
||||||
|
|
||||||
llama_model: "/path/to/llama2/weight"
|
llama_model: "please set this value to the path of llama2-chat-7b"
|
||||||
|
|
||||||
preprocess:
|
preprocess:
|
||||||
vis_processor:
|
vis_processor:
|
||||||
|
@ -15,7 +15,7 @@ model:
|
|||||||
# generation configs
|
# generation configs
|
||||||
prompt: ""
|
prompt: ""
|
||||||
|
|
||||||
llama_model: "/path/to/vicuna/weight"
|
llama_model: "please set this value to the path of vicuna model"
|
||||||
|
|
||||||
preprocess:
|
preprocess:
|
||||||
vis_processor:
|
vis_processor:
|
||||||
|
@ -11,7 +11,7 @@ model:
|
|||||||
# generation configs
|
# generation configs
|
||||||
prompt: ""
|
prompt: ""
|
||||||
|
|
||||||
llama_model: "/path/to/llama2/weight"
|
llama_model: "please set this value to the path of llama2-chat-7b"
|
||||||
lora_r: 64
|
lora_r: 64
|
||||||
lora_alpha: 16
|
lora_alpha: 16
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user