Update restart policy and latest results.
Browse files- app.py +2 -21
- latest_results.tsv +17 -17
app.py
CHANGED
@@ -35,19 +35,6 @@ def restart_space():
|
|
35 |
API.restart_space(repo_id=REPO_ID)
|
36 |
|
37 |
|
38 |
-
# try:
|
39 |
-
# print(EVAL_REQUESTS_PATH)
|
40 |
-
# snapshot_download(
|
41 |
-
# repo_id=QUEUE_REPO,
|
42 |
-
# local_dir=EVAL_REQUESTS_PATH,
|
43 |
-
# repo_type="dataset",
|
44 |
-
# tqdm_class=None,
|
45 |
-
# etag_timeout=30,
|
46 |
-
# token=TOKEN,
|
47 |
-
# )
|
48 |
-
# except Exception:
|
49 |
-
# restart_space()
|
50 |
-
|
51 |
try:
|
52 |
print("Saving results locally at:", EVAL_RESULTS_PATH)
|
53 |
snapshot_download(
|
@@ -65,12 +52,6 @@ except Exception:
|
|
65 |
raw_data, original_df = get_leaderboard_df(EVAL_RESULTS_PATH, EVAL_REQUESTS_PATH, COLS, BENCHMARK_COLS)
|
66 |
leaderboard_df = original_df.copy()
|
67 |
|
68 |
-
# (
|
69 |
-
# finished_eval_queue_df,
|
70 |
-
# running_eval_queue_df,
|
71 |
-
# pending_eval_queue_df,
|
72 |
-
# ) = get_evaluation_queue_df(EVAL_REQUESTS_PATH, EVAL_COLS)
|
73 |
-
|
74 |
|
75 |
# Searching and filtering
|
76 |
def update_table(
|
@@ -349,6 +330,6 @@ with demo:
|
|
349 |
)
|
350 |
|
351 |
scheduler = BackgroundScheduler()
|
352 |
-
scheduler.add_job(restart_space, "interval", seconds=
|
353 |
scheduler.start()
|
354 |
-
demo.queue(default_concurrency_limit=
|
|
|
35 |
API.restart_space(repo_id=REPO_ID)
|
36 |
|
37 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
38 |
try:
|
39 |
print("Saving results locally at:", EVAL_RESULTS_PATH)
|
40 |
snapshot_download(
|
|
|
52 |
raw_data, original_df = get_leaderboard_df(EVAL_RESULTS_PATH, EVAL_REQUESTS_PATH, COLS, BENCHMARK_COLS)
|
53 |
leaderboard_df = original_df.copy()
|
54 |
|
|
|
|
|
|
|
|
|
|
|
|
|
55 |
|
56 |
# Searching and filtering
|
57 |
def update_table(
|
|
|
330 |
)
|
331 |
|
332 |
scheduler = BackgroundScheduler()
|
333 |
+
scheduler.add_job(restart_space, "interval", seconds=7200)
|
334 |
scheduler.start()
|
335 |
+
demo.queue(default_concurrency_limit=20).launch()
|
latest_results.tsv
CHANGED
@@ -1,18 +1,18 @@
|
|
1 |
eval_name Precision Type T Weight type Architecture Model Average β¬οΈ Hub License #Params (B) Model sha Hub β€οΈ Available on the hub Code Data AMI 2020 Agg AMI 2020 Miso ARC-C Belebele GeNTE Neutralizing HaSpeeDe2 HS HaSpeeDe2 Stereo HateCheck HONEST IronITA Irony IronITA Sarcasm ItaCoLA News Sum SENTIPOLC SQuAD it TruthfulQA XCOPA
|
2 |
-
|
3 |
-
|
4 |
-
|
5 |
-
|
6 |
-
|
7 |
-
|
8 |
-
|
9 |
-
|
10 |
-
|
11 |
-
|
12 |
-
|
13 |
-
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
-
|
18 |
-
|
|
|
1 |
eval_name Precision Type T Weight type Architecture Model Average β¬οΈ Hub License #Params (B) Model sha Hub β€οΈ Available on the hub Code Data AMI 2020 Agg AMI 2020 Miso ARC-C Belebele GeNTE Neutralizing HaSpeeDe2 HS HaSpeeDe2 Stereo HateCheck HONEST IronITA Irony IronITA Sarcasm ItaCoLA News Sum SENTIPOLC SQuAD it TruthfulQA XCOPA
|
2 |
+
0 swap-uniba_LLaMAntino-3-ANITA-8B-Inst-DPO-ITA_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA</a>" 59.925692200303324 0 True ? ? 49.61934617107031 73.58604698054239 56.484641638225256 83.55555555555556 33.8255033557047 72.24399819126907 61.627116844508144 80.51511613552358 100.0 67.79529918401192 46.19514665929917 0.24261234404280246 33.783978293075634 46.49499761664646 71.27317142821833 68.09423700746308 73.4
|
3 |
+
9 mistralai_Mistral-7B-Instruct-v0.2_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mistralai/Mistral-7B-Instruct-v0.2</a>" 57.57154925481929 0 True ? ? 61.95096430524839 66.42194008585093 44.36860068259386 67.22222222222223 29.12751677852349 71.07491292799637 67.27017961567233 78.40873056250285 100.0 59.16469471738617 55.53851376330874 0.27708420363666786 36.377962201593874 50.02052664310759 68.04841543730598 59.24407318497844 64.2
|
4 |
+
2 mii-community_zefiro-7b-dpo-ITA_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-dpo-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-dpo-ITA</a>" 55.96825697198048 0 True ? ? 59.97920997920998 66.14027143881808 44.19795221843004 65.88888888888889 29.798657718120808 66.93068606112085 61.46209894750329 82.83622905315102 100.0 58.523449206965395 54.918191698733956 0.22337556862808253 35.66642647158017 38.80971929318383 74.34293876621986 43.34227321311386 68.4
|
5 |
+
7 meta-llama_Meta-Llama-3-8B_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/meta-llama/Meta-Llama-3-8B"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Meta-Llama-3-8B</a>" 55.933099551030125 0 True ? ? 60.02710027100271 63.14678395603251 40.529010238907844 76.0 29.53020134228188 65.30297764359561 59.541073390095356 79.81131536880565 100.0 57.31801541230962 56.750548188367965 0.2786244415689118 32.93607461627173 39.93136214294286 76.49082768675667 42.06877766857276 71.2
|
6 |
+
10 mii-community_zefiro-7b-base-ITA_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-base-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-base-ITA</a>" 55.286768709834995 0 True ? ? 60.14362403797995 64.54082375784897 40.955631399317404 58.55555555555556 28.456375838926174 66.12858980217781 59.74063711314884 82.46753086246828 100.0 59.05311714498798 57.8863223808017 0.09963712635854956 34.19887652648641 39.18986054178559 75.6692177776856 46.18926820166605 66.60000000000001
|
7 |
+
15 mii-community_zefiro-7b-sft-ITA_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-sft-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-sft-ITA</a>" 55.135348610310785 0 True ? ? 60.458679319889285 63.51256529535591 42.32081911262799 67.77777777777779 27.248322147651006 65.72752014372092 60.158604473839915 83.05031763559394 100.0 52.69566548195397 51.630329924754 0.08940878967203518 34.80608014621687 43.75098014181036 74.55382319645513 42.52003278796414 67.0
|
8 |
+
1 mistralai_Mistral-7B-v0.1_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/mistralai/Mistral-7B-v0.1"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mistralai/Mistral-7B-v0.1</a>" 54.550973703693096 0 True ? ? 60.52050697114497 63.66158365032981 41.21160409556314 65.66666666666666 29.53020134228188 60.38816689466484 57.907599364752336 80.59264657366079 100.0 55.23299236027556 55.67900219124808 0.131895692851752 34.09475870496535 38.87141003943634 75.08500650762954 43.19251190731156 65.60000000000001
|
9 |
+
14 swap-uniba_LLaMAntino-2-chat-13b-hf-ITA_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-chat-13b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-chat-13b-hf-ITA</a>" 53.88562700961127 0 True ? ? 61.41230947327803 64.77739009492042 39.07849829351536 60.44444444444444 25.503355704697988 67.1548291501024 59.101414060364085 81.83763297921335 100.0 57.92048929663609 52.2777996043644 0.1015435288181161 23.81691473597593 34.69232896418751 73.10003377486571 44.43667505800782 70.39999999999999
|
10 |
+
6 swap-uniba_LLaMAntino-2-13b-hf-ITA_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-13b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-13b-hf-ITA</a>" 51.26001015437534 0 True ? ? 56.79723502304148 60.93495016444478 38.56655290102389 52.33333333333333 24.697986577181208 57.1976786986929 54.2447910290625 68.16391542846057 100.0 56.51605280366516 51.571111501558086 0.16387751408972254 23.495330157527007 38.60258050721683 74.20709928774112 42.12767769734222 71.8
|
11 |
+
13 meta-llama_Llama-2-13b-hf_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/meta-llama/Llama-2-13b-hf"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Llama-2-13b-hf</a>" 51.16172039685661 0 True ? ? 53.24565637065637 59.32319654843206 39.93174061433447 49.666666666666664 24.295302013422816 54.13600451447075 54.88702987697715 74.1483219663718 100.0 50.34584608393744 49.636673785442774 0.11758183179468357 35.09699883531247 37.37259554778931 75.22840229480128 42.91722979615231 69.39999999999999
|
12 |
+
5 g8a9_tweety-mistral-7b_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/g8a9/tweety-mistral-7b"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">g8a9/tweety-mistral-7b</a>" 48.37800669811661 0 True ? ? 56.17170479302832 56.423255312264054 37.96928327645051 49.666666666666664 27.91946308724832 53.70259637851317 53.57434872305199 64.41588573083048 100.0 50.21506876304183 49.42973129711966 0.11006633622278786 18.81035591897043 28.46426204947685 64.39794432633592 37.75548120876122 73.4
|
13 |
+
4 meta-llama_Llama-2-7b-hf_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/meta-llama/Llama-2-7b-hf"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Llama-2-7b-hf</a>" 47.26821759114118 0 True ? ? 50.26836062232489 57.089775606014214 35.153583617747444 36.11111111111111 25.100671140939596 49.33536331841416 51.73318260900284 67.35406316275402 100.0 47.63910390674802 48.347086153434084 0.036528464070504335 33.756452251726735 27.82288694076669 68.6449557225095 39.16657442183614 66.0
|
14 |
+
8 swap-uniba_LLaMAntino-2-7b-hf-ITA_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-7b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-7b-hf-ITA</a>" 45.2518617559276 0 True ? ? 51.11111111111111 53.267951636107355 33.70307167235495 34.66666666666667 24.295302013422816 45.514286626950536 47.59019966407009 60.855425171736485 100.0 47.55193616643805 46.04838972288254 0.043130721156949686 24.582547279426233 22.260015178994326 69.30864535653794 40.48297086291322 68.0
|
15 |
+
12 sapienzanlp_Minerva-3B-base-v1.0_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-3B-base-v1.0</a>" 41.485187669928465 0 True ? ? 49.875480140137604 52.15633707230505 30.97269624573379 24.333333333333336 23.08724832214765 48.93622623624203 45.71528801169143 47.43110547988597 100.0 43.13118956315911 45.77114427860697 -0.015363788820154219 21.8700732759084 23.020245154283693 42.99174436502196 37.371442699146954 68.60000000000001
|
16 |
+
3 swap-uniba_LLaMAntino-2-chat-7b-hf-ITA_? ? ? Unknown LlamaForCausalLM "<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-chat-7b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-chat-7b-hf-ITA</a>" 38.997012161113425 0 True ? ? 47.32809806550469 43.776841477788466 29.180887372013654 28.111111111111107 23.48993288590604 41.57668822526659 41.556830771361305 44.984357634264406 100.0 41.716872329343005 43.53102430893341 -0.02574637563194932 8.269309204888462 9.339380225529704 58.43272201840739 39.880897484241935 61.8
|
17 |
+
16 sapienzanlp_Minerva-1B-base-v1.0_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-1B-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-1B-base-v1.0</a>" 38.906733116823304 0 True ? ? 50.76172656624852 53.84641914146224 24.573378839590443 22.666666666666664 26.57718120805369 48.25128927047713 44.581537708222804 50.10425395808837 100.0 46.49541549308013 45.46046920890855 0.022249590030925144 14.27287574762189 16.571464690513597 17.48160254077023 39.747932356260876 60.0
|
18 |
+
11 sapienzanlp_Minerva-350M-base-v1.0_? ? ? Unknown MistralForCausalLM "<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-350M-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-350M-base-v1.0</a>" 36.95204565967993 0 True ? ? 45.17543859649123 35.72145622912868 24.40273037542662 22.88888888888889 52.75167785234899 41.92832319168979 40.67042217927179 46.277755136438564 100.0 36.23277134884009 43.223117993157416 -0.036868413829916326 10.308018221966565 23.388373345290127 4.903980027793706 43.7486912416563 56.599999999999994
|