#!/bin/bash python -m llava.eval.model_vqa \ --model-path liuhaotian/llava-v1.5-13b \ --question-file ./playground/data/eval/llava-bench-in-the-wild/questions.jsonl \ --image-folder ./playground/data/eval/llava-bench-in-the-wild/images \ --answers-file ./playground/data/eval/llava-bench-in-the-wild/answers/llava-v1.5-13b.jsonl \ --temperature 0 \ --conv-mode vicuna_v1 mkdir -p playground/data/eval/llava-bench-in-the-wild/reviews python llava/eval/eval_gpt_review_bench.py \ --question playground/data/eval/llava-bench-in-the-wild/questions.jsonl \ --context playground/data/eval/llava-bench-in-the-wild/context.jsonl \ --rule llava/eval/table/rule.json \ --answer-list \ playground/data/eval/llava-bench-in-the-wild/answers_gpt4.jsonl \ playground/data/eval/llava-bench-in-the-wild/answers/llava-v1.5-13b.jsonl \ --output \ playground/data/eval/llava-bench-in-the-wild/reviews/llava-v1.5-13b.jsonl python llava/eval/summarize_gpt_review.py -f playground/data/eval/llava-bench-in-the-wild/reviews/llava-v1.5-13b.jsonl