teowu commited on
Commit
1a17291
·
1 Parent(s): 5e9cb18

add A2/A3 existing results

Browse files
Files changed (3) hide show
  1. app.py +7 -2
  2. qbench_a2_single.csv +22 -0
  3. qbench_a3_single.csv +23 -0
app.py CHANGED
@@ -30,9 +30,14 @@ with block:
30
  gr.Markdown(
31
  LEADERBORAD_INTRODUCTION
32
  )
33
- with gr.Tab("Perception-A1-dev"):
34
  gr.DataFrame(pd.read_csv("qbench_a1_single_dev.csv"))
35
- with gr.Tab("Perception-A1-test"):
36
  gr.DataFrame(pd.read_csv("qbench_a1_single_test.csv"))
 
 
 
 
 
37
 
38
  block.launch(share=True)
 
30
  gr.Markdown(
31
  LEADERBORAD_INTRODUCTION
32
  )
33
+ with gr.Tab("Perception (A1, dev)"):
34
  gr.DataFrame(pd.read_csv("qbench_a1_single_dev.csv"))
35
+ with gr.Tab("Perception (A1,test)"):
36
  gr.DataFrame(pd.read_csv("qbench_a1_single_test.csv"))
37
+ with gr.Tab("Description (A2)"):
38
+ gr.DataFrame(pd.read_csv("qbench_a2_single.csv"))
39
+ with gr.Tab("Assessment (A3)"):
40
+ gr.DataFrame(pd.read_csv("qbench_a3_single.csv"))
41
+
42
 
43
  block.launch(share=True)
qbench_a2_single.csv ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model,Completeness,Precision,Relevance,Sum
2
+ InfiMM (Zephyr-7B),0.77,1.08,1.71,3.58
3
+ Emu2-Chat (LLaMA-33B),1.07,1.24,1.88,4.19
4
+ Fuyu-8B (Persimmon-8B),0.88,0.83,1.82,3.53
5
+ BakLLava (Mistral-7B),1.0,0.77,1.61,3.38
6
+ SPHINX,0.79,1.14,1.72,3.65
7
+ mPLUG-Owl2 (LLaMA-7B),1.06,1.24,1.36,3.67
8
+ LLaVA-v1.5 (Vicuna-v1.5-7B),0.9,1.13,1.18,3.21
9
+ LLaVA-v1.5 (Vicuna-v1.5-13B),0.91,1.28,1.29,3.47
10
+ InternLM-XComposer-VL (InternLM),1.08,1.26,1.87,4.21
11
+ IDEFICS-Instruct (LLaMA-7B),0.83,1.03,1.33,3.18
12
+ Qwen-VL (QwenLM),0.98,0.75,1.63,3.36
13
+ Shikra (Vicuna-7B),0.89,1.11,1.33,3.34
14
+ Otter-v1 (MPT-7B),0.96,0.83,1.83,3.61
15
+ Kosmos-2,1.12,1.06,1.85,4.03
16
+ InstructBLIP (Flan-T5-XL),0.87,1.04,1.11,3.02
17
+ InstructBLIP (Vicuna-7B),0.79,1.21,0.84,2.84
18
+ VisualGLM-6B (GLM-6B),0.82,0.97,1.21,2.99
19
+ mPLUG-Owl (LLaMA-7B),1.06,1.28,1.6,3.94
20
+ LLaMA-Adapter-V2,0.85,1.15,1.44,3.45
21
+ LLaVA-v1 (Vicuna-13B),0.91,1.25,1.6,3.76
22
+ MiniGPT-4 (Vicuna-13B),1.0,1.26,1.41,3.67
qbench_a3_single.csv ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Model,KONiQ-10k,SPAQ,LIVE-FB,LIVE-ITW,CGIQA-6K,AGIQA-3K
2
+ NIQE,0.316/0.377,0.693/0.669,0.211/0.288,0.480/0.451,0.075/0.056,0.562/0.517
3
+ CLIP-ViT-Large-14,0.468/0.505,0.385/0.389,0.218/0.237,0.307/0.308,0.285/0.290,0.436/0.458
4
+ InfiMM (Zephyr-7B),0.507/0.547,0.616/0.633,0.269/0.299,0.548/0.580,0.229/0.245,0.706/0.767
5
+ Emu2-Chat (LLaMA-33B),0.664/0.714,0.712/0.698,0.355/0.341,0.597/0.611,0.224/0.269,0.759/0.751
6
+ Fuyu-8B (Persimmon-8B),0.124/0.123,0.125/0.179,0.164/0.133,0.225/0.176,0.118/0.116,0.368/0.317
7
+ BakLLava (Mistral-7B),0.389/0.390,0.406/0.398,0.227/0.216,0.335/0.337,0.179/0.209,0.542/0.561
8
+ mPLUG-Owl2 (LLaMA-7B),0.196/0.252,0.589/0.614,0.217/0.286,0.293/0.342,-0.024/-0.032,0.473/0.492
9
+ LLaVA-v1.5 (Vicuna-v1.5-7B),0.463/0.459,0.443/0.467,0.305/0.321,0.344/0.358,0.321/0.333,0.672/0.738
10
+ LLaVA-v1.5 (Vicuna-v1.5-13B),0.448/0.460,0.563/0.584,0.310/0.339,0.445/0.481,0.285/0.297,0.664/0.754
11
+ InternLM-XComposer-VL (InternLM),0.564/0.615,0.730/0.750,0.360/0.416,0.612/0.676,0.243/0.265,0.732/0.775
12
+ IDEFICS-Instruct (LLaMA-7B),0.375/0.400,0.474/0.484,0.235/0.240,0.409/0.428,0.244/0.227,0.562/0.622
13
+ Qwen-VL (QwenLM),0.470/0.546,0.676/0.669,0.298/0.338,0.504/0.532,0.273/0.284,0.617/0.686
14
+ Shikra (Vicuna-7B),0.314/0.307,0.320/0.337,0.237/0.241,0.322/0.336,0.198/0.201,0.640/0.661
15
+ Otter-v1 (MPT-7B),0.406/0.406,0.436/0.441,0.143/0.142,-0.008/0.018,0.254/0.264,0.475/0.481
16
+ Kosmos-2,0.255/0.281,0.644/0.641,0.196/0.195,0.358/0.368,0.210/0.225,0.489/0.491
17
+ InstructBLIP (Flan-T5-XL),0.334/0.362,0.582/0.599,0.248/0.267,0.113/0.113,0.167/0.188,0.378/0.400
18
+ InstructBLIP (Vicuna-7B),0.359/0.437,0.683/0.689,0.200/0.283,0.253/0.367,0.263/0.304,0.629/0.663
19
+ VisualGLM-6B (GLM-6B),0.247/0.234,0.498/0.507,0.146/0.154,0.110/0.116,0.209/0.183,0.342/0.349
20
+ mPLUG-Owl (LLaMA-7B),0.409/0.427,0.634/0.644,0.241/0.271,0.437/0.487,0.148/0.180,0.687/0.711
21
+ LLaMA-Adapter-V2,0.354/0.363,0.464/0.506,0.275/0.329,0.298/0.360,0.257/0.271,0.604/0.666
22
+ LLaVA-v1 (Vicuna-13B),0.462/0.457,0.442/0.462,0.264/0.280,0.404/0.417,0.208/0.237,0.626/0.684
23
+ MiniGPT-4 (Vicuna-13B),0.239/0.257,0.238/0.253,0.170/0.183,0.339/0.340,0.252/0.246,0.572/0.591