By default, this leaderboard is sorted by overall results. To view other sorted results, please click on the corresponding cell. Colored rows indicate closed-source models/APIs.
# | Model | LLM Params |
Date | Overall (%) | Perception (%) | Reasoning (%) | Probing (%) |
---|---|---|---|---|---|---|---|
1 |
InternVL2-40B
Shanghai AI Lab |
40B | 2024/10/01 | 67.95 | 65.44 | 73.99 | 59.59 |
2 | InternVL2-76B
Shanghai AI Lab |
76B | 2024/10/01 | 67.28 | 63.41 | 75.44 | 58.46 |
3 | Qwen2-VL-72B
Alibaba |
72B | 2024/10/01 | 65.24 | 56.53 | 76.39 | 70.26 |
4 |
InternVL-Chat-V1.2-Plus
Shanghai AI Lab |
40B | 2024/10/01 | 64.94 | 60.73 | 70.78 | 65.80 |
5 |
InternVL2.5-78B
Shanghai AI Lab |
78B | 2024/12/15 | 64.94 | 61.16 | 73.80 | 37.58 |
6 |
InternVL2-26B
Shanghai AI Lab |
26B | 2024/10/01 | 63.08 | 60.40 | 70.03 | 52.43 |
7 |
VILA1.5-40B
NVIDIA & MIT |
40B | 2024/10/01 | 63.08 | 60.40 | 70.03 | 52.43 |
8 | Qwen2-VL-7B
Alibaba |
7B | 2024/10/01 | 62.69 | 60.22 | 72.42 | 41.43 |
9 |
InternVL2.5-38B
Shanghai AI Lab |
38B | 2024/12/15 | 62.21 | 59.13 | 72.61 | 41.68 |
10 |
InternVL2.5-26B
Shanghai AI Lab |
26B | 2024/12/15 | 61.33 | 58.01 | 72.42 | 39.58 |
11 |
Ovis1.6-Gemma2-27B
Alibaba |
27B | 2024/12/15 | 60.88 | 56.98 | 72.23 | 41.14 |
12 |
Molmo-72B
Allen AI |
72B | 2024/12/15 | 60.16 | 61.52 | 63.60 | 41.47 |
13 | GPT-4o
OpenAI |
- | 2024/10/01 | 59.71 | 57.63 | 64.17 | 54.65 |
14 | InternVL-Chat-v1.2
Shanghai AI Lab |
40B | 2024/10/01 | 59.61 | 56.49 | 63.79 | 60.71 |
15 | InternVL-Chat-v1.5
Shanghai AI Lab |
26B | 2024/10/01 | 59.40 | 53.68 | 68.20 | 57.01 |
16 |
POINTS1.5-Qwen2.5-7B-Chat
WeChat AI |
7B | 2024/12/15 | 59.37 | 56.31 | 67.88 | 45.21 |
17 |
Ovis1.6-Gemma2-9B
Alibaba |
9B | 2024/12/15 | 58.52 | 53.75 | 71.16 | 38.63 |
18 | InternVL2-8B
Shanghai AI Lab |
8B | 2024/10/01 | 58.47 | 53.44 | 67.00 | 54.10 |
19 | LLaVA-1.6-34B
NTU & UW Madison & ByteDance |
34B | 2024/10/01 | 58.25 | 57.82 | 58.88 | 58.17 |
20 |
Llama-3.2-90B-Vision
Meta |
90B | 2024/12/15 | 57.92 | 58.26 | 60.64 | 46.84 |
21 |
InternVL2.5-8B
Shanghai AI Lab |
8B | 2024/12/15 | 57.54 | 53.68 | 68.07 | 40.41 |
22 | MiniCPM-V2.6
Tsinghua University |
8B | 2024/10/01 | 57.01 | 55.36 | 60.14 | 54.43 |
23 | InternLM-XComposer2-4KHD-7B
Shanghai AI Lab & CUHK & SenseTime |
7B | 2024/10/01 | 56.69 | 52.55 | 61.71 | 60.02 |
24 |
InternVL2.5-4B
Shanghai AI Lab |
4B | 2024/12/15 | 55.44 | 52.60 | 63.66 | 41.21 |
25 | Qwen-VL-Max
Alibaba |
- | 2024/10/01 | 55.18 | 50.36 | 59.63 | 63.87 |
26 | InternLM-XComposer2.5-7B
Shanghai AI Lab & CUHK & SenseTime |
7B | 2024/10/01 | 55.10 | 50.61 | 63.16 | 49.64 |
27 | Hunyuan-Vision
Tencent |
- | 2024/10/01 | 54.64 | 54.80 | 57.18 | 45.03 |
28 | InternLM-Xcomposer2-VL
Shanghai AI Lab & CUHK & SenseTime |
7B | 2024/10/01 | 54.62 | 51.25 | 58.75 | 57.15 |
29 | Qwen2-VL-2B
Alibaba |
2B | 2024/10/01 | 53.78 | 50.21 | 61.52 | 44.81 |
30 | Gemini-1.5-Pro
|
- | 2024/10/01 | 53.27 | 50.64 | 58.12 | 49.60 |
31 | Mini-Gemini-34B
CUHK & SmartMore |
34B | 2024/10/01 | 53.06 | 51.25 | 58.94 | 41.79 |
32 |
Ovis1.6-Llama3.2-3B
Alibaba |
3B | 2024/12/15 | 52.78 | 50.12 | 62.41 | 32.79 |
33 |
Molmo-7B-D
Allen AI |
7B | 2024/12/15 | 52.61 | 52.12 | 56.17 | 42.70 |
34 | InternVL2-4B
Shanghai AI Lab |
4B | 2024/10/01 | 52.03 | 46.94 | 62.53 | 41.18 |
35 | LLaMA-3.2-11B-Vision-Instruct
Meta |
11B | 2024/10/01 | 52.01 | 50.88 | 54.47 | 49.17 |
36 |
InternVL2.5-2B
Shanghai AI Lab |
2B | 2024/12/15 | 51.62 | 50.36 | 56.86 | 39.80 |
37 | MiniCPM-Llama3-V2.5
Tsinghua University |
8B | 2024/10/01 | 51.54 | 45.68 | 62.85 | 41.79 |
38 | Mini-Gemini-34B-HD
CUHK & SmartMore |
34B | 2024/10/01 | 51.48 | 47.73 | 61.40 | 35.91 |
39 | Bunny-LLaMA-3-V
BAAI |
8B | 2024/10/01 | 50.81 | 47.81 | 52.64 | 59.44 |
40 | Mini-Monkey
HUST |
2B | 2024/10/01 | 50.41 | 47.81 | 56.49 | 42.37 |
41 | Phi3.5-Vision-Instruct
Microsoft |
4.2B | 2024/10/01 | 50.02 | 45.97 | 54.53 | 54.65 |
42 | ColgVLM2-Llama3-Chat-19B
Zhipu AI |
19B | 2024/10/01 | 49.84 | 50.34 | 48.87 | 50.69 |
43 | Phi3-Vision-Instruct
Microsoft |
4.2B | 2024/10/01 | 48.52 | 45.55 | 50.44 | 56.75 |
44 | Yi-VL-34B
01.AI |
34B | 2024/10/01 | 47.86 | 42.99 | 53.15 | 53.88 |
45 | Step-1V-32K
Stepfun |
- | 2024/10/01 | 47.64 | 41.25 | 57.49 | 45.46 |
46 | ConvLLaVA-1024-7B
Alibaba & Tsinghua University |
7B | 2024/10/01 | 47.32 | 43.70 | 54.41 | 40.89 |
47 | Yi-VL-6B
01.AI |
6B | 2024/10/01 | 46.87 | 43.80 | 50.76 | 48.76 |
48 | Bunny-3B
BAAI |
3B | 2024/10/01 | 46.32 | 43.42 | 47.98 | 55.08 |
49 | Bunny-4B-V1.0
BAAI |
3B | 2024/10/01 | 46.07 | 43.68 | 50.50 | 42.66 |
50 | LLaVA-HR-13B
Xiamen University |
13B | 2024/10/01 | 46.02 | 41.83 | 51.26 | 48.80 |
51 | ConvLLaVA-1536-7B
Alibaba & Tsinghua University |
7B | 2024/10/01 | 45.52 | 41.84 | 54.09 | 34.20 |
52 | InternVL2-2B
Shanghai AI Lab |
2B | 2024/10/01 | 45.11 | 42.37 | 51.07 | 38.16 |
53 | Monkey-Chat
HUST |
7.7B | 2024/10/01 | 44.90 | 41.79 | 48.24 | 48.91 |
54 |
InternVL2.5-1B
Shanghai AI Lab |
1B | 2024/12/15 | 44.76 | 43.45 | 48.93 | 36.93 |
55 |
Molmo-7B-O
Allen AI |
7B | 2024/12/15 | 44.02 | 42.89 | 49.24 | 31.59 |
56 | Mini-Gemini-13B
CUHK & SmartMore |
13B | 2024/10/01 | 43.74 | 38.51 | 54.60 | 32.28 |
57 | SliME-7B
UCAS & Squirrel AI & Alibaba & Meta |
7B | 2024/10/01 | 43.45 | 40.56 | 51.51 | 30.03 |
58 | INF-LLaVA*
Xiamen University |
8B | 2024/10/01 | 43.32 | 40.13 | 51.39 | 31.41 |
59 | SliME-8B
UCAS & Squirrel AI & Alibaba & Meta |
8B | 2024/10/01 | 43.29 | 40.44 | 51.26 | 29.96 |
60 | INF-LLaVA
Xiamen University |
8B | 2024/10/01 | 43.04 | 41.80 | 46.98 | 35.58 |
61 | LLaVA-HR-7B
Xiamen University |
7B | 2024/10/01 | 42.73 | 39.38 | 50.38 | 33.04 |
62 | SliME-13B
UCAS & Squirrel AI & Alibaba & Meta |
13B | 2024/10/01 | 42.63 | 39.30 | 50.06 | 33.55 |
63 | ConvLLaVA-768-7B
Alibaba & Tsinghua University |
7B | 2024/10/01 | 42.40 | 36.51 | 52.46 | 37.11 |
64 | InternVL2-1B
Shanghai AI Lab |
1B | 2024/10/01 | 42.06 | 39.65 | 49.62 | 27.89 |
65 | Mini-Gemini-13B-HD
CUHK & SmartMore |
13B | 2024/10/01 | 41.99 | 37.24 | 51.07 | 34.28 |
66 | Qwen-VL-Chat
Alibaba |
13B | 2024/10/01 | 41.64 | 36.10 | 49.69 | 41.54 |
67 | DeepStack-L-HD-Vicuna-7B
Fudan University & Microsoft |
7B | 2024/10/01 | 40.26 | 35.19 | 48.87 | 35.88 |
68 | DeepStack-L-Vicuna-7B
Fudan University & Microsoft |
7B | 2024/10/01 | 39.75 | 36.92 | 46.60 | 30.21 |
69 | mPLUG-Owl2-7B
Alibaba |
7B | 2024/10/01 | 39.59 | 36.90 | 46.16 | 30.36 |
70 | LLaVA-1.6-Vicuna-13B
NTU & UW Madison & ByteDance |
13B | 2024/10/01 | 38.03 | 31.15 | 47.92 | 38.16 |
71 | LLaVA-1.6-Mistral-7B
NTU & UW Madison & ByteDance |
7B | 2024/10/01 | 37.18 | 33.64 | 42.00 | 38.24 |
72 | InstructBLIP-Vicuna-13B
Salesforce |
13B | 2024/10/01 | 37.06 | 35.53 | 42.70 | 25.24 |
73 | InstructBLIP-Vicuna-7B
Salesforce |
7B | 2024/10/01 | 36.86 | 33.22 | 43.70 | 31.41 |
74 | LLaVA-1.5-13B
NTU & UW Madison & ByteDance |
13B | 2024/10/01 | 36.07 | 29.91 | 43.45 | 41.39 |
75 | Random Choice
- |
- | 2024/10/01 | 30.15 | 24.88 | 38.22 | 28.61 |