- [2013-12-28]:Model and inference released.
- Because QQMM use lora tuning the LLM. Before evaluation, you should download the base LLM model:
vicuna-13b-v1.5 - QQMM uses visual encoder. Please download from openai/clip-vit-large-patch14-336
Before evaluation, please download the MME benchmark files MME into mme_bench directory.
Then run the cmd:
sh eval/eval_mme.sh
QQMM achieved xxx points, which was topx on MME benchmark at 2023-12-28.