ZebangCheng's picture
eval configs
292d4e3
raw
history blame
537 Bytes
model:
arch: minigpt_v2
model_type: pretrain
max_txt_len: 500
end_sym: "</s>"
# low_resource: True
low_resource: False
prompt_template: '[INST] {} [/INST]'
ckpt: "/home/czb/project/Emotion-LLaMA/checkpoints/save_checkpoint/MER2023-20240602194_demo/Emoation_LLaMA.pth"
lora_r: 64
lora_alpha: 16
datasets:
feature_face_caption:
vis_processor:
train:
name: "blip2_image_eval"
image_size: 448
text_processor:
train:
name: "blip_caption"
run:
task: image_text_pretrain