Oskar Douwe van der Wal
commited on
Commit
·
7bc099d
1
Parent(s):
52534dd
New results
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- pythia-14m-seed0/step0/EleutherAI__pythia-14m/results_2024-08-15T12-48-54.868316.json +96 -0
- pythia-14m-seed0/step1/EleutherAI__pythia-14m/results_2024-08-15T12-50-21.093189.json +96 -0
- pythia-14m-seed0/step1000/EleutherAI__pythia-14m/results_2024-08-15T13-04-58.314501.json +96 -0
- pythia-14m-seed0/step10000/EleutherAI__pythia-14m/results_2024-08-15T13-18-05.667000.json +96 -0
- pythia-14m-seed0/step100000/EleutherAI__pythia-14m/results_2024-08-15T15-27-54.136872.json +96 -0
- pythia-14m-seed0/step101000/EleutherAI__pythia-14m/results_2024-08-15T15-29-20.379855.json +96 -0
- pythia-14m-seed0/step102000/EleutherAI__pythia-14m/results_2024-08-15T15-30-45.504373.json +96 -0
- pythia-14m-seed0/step103000/EleutherAI__pythia-14m/results_2024-08-15T15-32-11.742196.json +96 -0
- pythia-14m-seed0/step104000/EleutherAI__pythia-14m/results_2024-08-15T15-33-38.569825.json +96 -0
- pythia-14m-seed0/step105000/EleutherAI__pythia-14m/results_2024-08-15T15-35-05.211421.json +96 -0
- pythia-14m-seed0/step106000/EleutherAI__pythia-14m/results_2024-08-15T15-36-32.465444.json +96 -0
- pythia-14m-seed0/step107000/EleutherAI__pythia-14m/results_2024-08-15T15-38-00.033086.json +96 -0
- pythia-14m-seed0/step108000/EleutherAI__pythia-14m/results_2024-08-15T15-39-28.787439.json +96 -0
- pythia-14m-seed0/step109000/EleutherAI__pythia-14m/results_2024-08-15T15-40-55.857604.json +96 -0
- pythia-14m-seed0/step11000/EleutherAI__pythia-14m/results_2024-08-15T13-19-30.603325.json +96 -0
- pythia-14m-seed0/step110000/EleutherAI__pythia-14m/results_2024-08-15T15-42-21.675993.json +96 -0
- pythia-14m-seed0/step111000/EleutherAI__pythia-14m/results_2024-08-15T15-43-46.617900.json +96 -0
- pythia-14m-seed0/step112000/EleutherAI__pythia-14m/results_2024-08-15T15-45-12.514667.json +96 -0
- pythia-14m-seed0/step113000/EleutherAI__pythia-14m/results_2024-08-15T15-46-38.770492.json +96 -0
- pythia-14m-seed0/step114000/EleutherAI__pythia-14m/results_2024-08-15T15-48-05.260016.json +96 -0
- pythia-14m-seed0/step115000/EleutherAI__pythia-14m/results_2024-08-15T15-49-32.634482.json +96 -0
- pythia-14m-seed0/step116000/EleutherAI__pythia-14m/results_2024-08-15T15-51-01.124488.json +96 -0
- pythia-14m-seed0/step117000/EleutherAI__pythia-14m/results_2024-08-15T15-52-30.240629.json +96 -0
- pythia-14m-seed0/step118000/EleutherAI__pythia-14m/results_2024-08-15T15-53-56.953504.json +96 -0
- pythia-14m-seed0/step119000/EleutherAI__pythia-14m/results_2024-08-15T15-55-23.128925.json +96 -0
- pythia-14m-seed0/step12000/EleutherAI__pythia-14m/results_2024-08-15T13-20-56.302395.json +96 -0
- pythia-14m-seed0/step120000/EleutherAI__pythia-14m/results_2024-08-15T15-56-52.216443.json +96 -0
- pythia-14m-seed0/step121000/EleutherAI__pythia-14m/results_2024-08-15T15-58-17.456909.json +96 -0
- pythia-14m-seed0/step122000/EleutherAI__pythia-14m/results_2024-08-15T15-59-44.315496.json +96 -0
- pythia-14m-seed0/step123000/EleutherAI__pythia-14m/results_2024-08-15T16-01-10.335351.json +96 -0
- pythia-14m-seed0/step124000/EleutherAI__pythia-14m/results_2024-08-15T16-02-35.938574.json +96 -0
- pythia-14m-seed0/step125000/EleutherAI__pythia-14m/results_2024-08-15T16-04-02.754305.json +96 -0
- pythia-14m-seed0/step126000/EleutherAI__pythia-14m/results_2024-08-15T16-05-29.292044.json +96 -0
- pythia-14m-seed0/step127000/EleutherAI__pythia-14m/results_2024-08-15T16-06-56.433830.json +96 -0
- pythia-14m-seed0/step128/EleutherAI__pythia-14m/results_2024-08-15T13-00-34.735877.json +96 -0
- pythia-14m-seed0/step128000/EleutherAI__pythia-14m/results_2024-08-15T16-08-22.902858.json +96 -0
- pythia-14m-seed0/step129000/EleutherAI__pythia-14m/results_2024-08-15T16-09-48.457182.json +96 -0
- pythia-14m-seed0/step13000/EleutherAI__pythia-14m/results_2024-08-15T13-22-24.295538.json +96 -0
- pythia-14m-seed0/step130000/EleutherAI__pythia-14m/results_2024-08-15T16-11-14.614730.json +96 -0
- pythia-14m-seed0/step131000/EleutherAI__pythia-14m/results_2024-08-15T16-12-40.692551.json +96 -0
- pythia-14m-seed0/step132000/EleutherAI__pythia-14m/results_2024-08-15T16-14-06.187317.json +96 -0
- pythia-14m-seed0/step133000/EleutherAI__pythia-14m/results_2024-08-15T16-15-33.241417.json +96 -0
- pythia-14m-seed0/step134000/EleutherAI__pythia-14m/results_2024-08-15T16-17-00.743542.json +96 -0
- pythia-14m-seed0/step135000/EleutherAI__pythia-14m/results_2024-08-15T16-18-28.514020.json +96 -0
- pythia-14m-seed0/step136000/EleutherAI__pythia-14m/results_2024-08-15T16-19-54.399480.json +96 -0
- pythia-14m-seed0/step137000/EleutherAI__pythia-14m/results_2024-08-15T16-21-20.940162.json +96 -0
- pythia-14m-seed0/step138000/EleutherAI__pythia-14m/results_2024-08-15T16-22-46.483201.json +96 -0
- pythia-14m-seed0/step139000/EleutherAI__pythia-14m/results_2024-08-15T16-24-12.061095.json +96 -0
- pythia-14m-seed0/step14000/EleutherAI__pythia-14m/results_2024-08-15T13-23-49.471021.json +96 -0
- pythia-14m-seed0/step140000/EleutherAI__pythia-14m/results_2024-08-15T16-25-37.975454.json +96 -0
pythia-14m-seed0/step0/EleutherAI__pythia-14m/results_2024-08-15T12-48-54.868316.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 3507781.8105202965,
|
5 |
+
"perplexity_stderr,none": 338907.2426000304,
|
6 |
+
"acc,none": 0.0,
|
7 |
+
"acc_stderr,none": 0.0,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step0,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step0",
|
71 |
+
"model_sha": "bc249c50b4a381f99e9296f08a69b8562a31fde4",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723751264.2585666,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1133831.651542538,
|
94 |
+
"end_time": 1133915.107703607,
|
95 |
+
"total_evaluation_time_seconds": "83.45616106898524"
|
96 |
+
}
|
pythia-14m-seed0/step1/EleutherAI__pythia-14m/results_2024-08-15T12-50-21.093189.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 3507781.8105202965,
|
5 |
+
"perplexity_stderr,none": 338907.2426000304,
|
6 |
+
"acc,none": 0.0,
|
7 |
+
"acc_stderr,none": 0.0,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step1,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step1",
|
71 |
+
"model_sha": "71caa8178c55d927b2234a30bbabac6bd072cae8",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723751353.851652,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1133922.894150664,
|
94 |
+
"end_time": 1134001.333325669,
|
95 |
+
"total_evaluation_time_seconds": "78.43917500483803"
|
96 |
+
}
|
pythia-14m-seed0/step1000/EleutherAI__pythia-14m/results_2024-08-15T13-04-58.314501.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 195988.90526637834,
|
5 |
+
"perplexity_stderr,none": 12193.7268807034,
|
6 |
+
"acc,none": 0.0,
|
7 |
+
"acc_stderr,none": 0.0,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step1000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step1000",
|
71 |
+
"model_sha": "67ed4908dc8d91741d246c4d9fbfa03e78c8505d",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723752224.1879683,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1134793.204587141,
|
94 |
+
"end_time": 1134878.555076991,
|
95 |
+
"total_evaluation_time_seconds": "85.35048984992318"
|
96 |
+
}
|
pythia-14m-seed0/step10000/EleutherAI__pythia-14m/results_2024-08-15T13-18-05.667000.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1511.251720455259,
|
5 |
+
"perplexity_stderr,none": 70.29803394374686,
|
6 |
+
"acc,none": 0.1003299049097613,
|
7 |
+
"acc_stderr,none": 0.004185707878554855,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step10000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step10000",
|
71 |
+
"model_sha": "50c427ff77acdb2553ee434ffabe182f5be3c1e2",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723753014.309483,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1135582.228500084,
|
94 |
+
"end_time": 1135665.902803599,
|
95 |
+
"total_evaluation_time_seconds": "83.67430351488292"
|
96 |
+
}
|
pythia-14m-seed0/step100000/EleutherAI__pythia-14m/results_2024-08-15T15-27-54.136872.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2480.572303985594,
|
5 |
+
"perplexity_stderr,none": 110.31519283833292,
|
6 |
+
"acc,none": 0.05142635358043858,
|
7 |
+
"acc_stderr,none": 0.0030770918924100804,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step100000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step100000",
|
71 |
+
"model_sha": "c5ab9636d47fcd482787a66ae2ef8a71002648b9",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723760806.4035344,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143373.76838171,
|
94 |
+
"end_time": 1143454.366109349,
|
95 |
+
"total_evaluation_time_seconds": "80.59772763913497"
|
96 |
+
}
|
pythia-14m-seed0/step101000/EleutherAI__pythia-14m/results_2024-08-15T15-29-20.379855.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1936.7673989556679,
|
5 |
+
"perplexity_stderr,none": 89.14269698503178,
|
6 |
+
"acc,none": 0.07354938870560838,
|
7 |
+
"acc_stderr,none": 0.0036367440978540603,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step101000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step101000",
|
71 |
+
"model_sha": "2a340986858902da1c8563a6c273066aaccc8750",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723760893.3566766,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143461.457184083,
|
94 |
+
"end_time": 1143540.619783163,
|
95 |
+
"total_evaluation_time_seconds": "79.16259908000939"
|
96 |
+
}
|
pythia-14m-seed0/step102000/EleutherAI__pythia-14m/results_2024-08-15T15-30-45.504373.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2332.1576555643323,
|
5 |
+
"perplexity_stderr,none": 103.2466067379606,
|
6 |
+
"acc,none": 0.055695711236173104,
|
7 |
+
"acc_stderr,none": 0.0031950593588073433,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step102000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step102000",
|
71 |
+
"model_sha": "e1ef05f9ceddbf4c8b5eda7b6e1eb66da607dc76",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723760978.5723045,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143547.70613283,
|
94 |
+
"end_time": 1143625.744282615,
|
95 |
+
"total_evaluation_time_seconds": "78.0381497850176"
|
96 |
+
}
|
pythia-14m-seed0/step103000/EleutherAI__pythia-14m/results_2024-08-15T15-32-11.742196.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2799.559936979982,
|
5 |
+
"perplexity_stderr,none": 123.34505192306801,
|
6 |
+
"acc,none": 0.05491946438967592,
|
7 |
+
"acc_stderr,none": 0.0031740197788982893,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step103000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step103000",
|
71 |
+
"model_sha": "2f4b2967a8d878296de9228ba494cd18a59fe35c",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761064.4161794,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143633.446766386,
|
94 |
+
"end_time": 1143711.9823575,
|
95 |
+
"total_evaluation_time_seconds": "78.53559111384675"
|
96 |
+
}
|
pythia-14m-seed0/step104000/EleutherAI__pythia-14m/results_2024-08-15T15-33-38.569825.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2003.2493640780494,
|
5 |
+
"perplexity_stderr,none": 89.98672601843188,
|
6 |
+
"acc,none": 0.06617504366388512,
|
7 |
+
"acc_stderr,none": 0.003463314484220094,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step104000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step104000",
|
71 |
+
"model_sha": "678191556681a68fbcd641ab2bd018f3634fd545",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761150.9524848,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143719.967749508,
|
94 |
+
"end_time": 1143798.810041786,
|
95 |
+
"total_evaluation_time_seconds": "78.8422922778409"
|
96 |
+
}
|
pythia-14m-seed0/step105000/EleutherAI__pythia-14m/results_2024-08-15T15-35-05.211421.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2204.366599640772,
|
5 |
+
"perplexity_stderr,none": 101.36401689274162,
|
6 |
+
"acc,none": 0.06966815447312245,
|
7 |
+
"acc_stderr,none": 0.003546893672151897,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step105000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step105000",
|
71 |
+
"model_sha": "52a764e3d099f01bcf6079282458009fba8e9815",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761237.5477817,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143806.663444182,
|
94 |
+
"end_time": 1143885.451216583,
|
95 |
+
"total_evaluation_time_seconds": "78.78777240100317"
|
96 |
+
}
|
pythia-14m-seed0/step106000/EleutherAI__pythia-14m/results_2024-08-15T15-36-32.465444.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2260.64934437262,
|
5 |
+
"perplexity_stderr,none": 101.92424646136189,
|
6 |
+
"acc,none": 0.06986221618474675,
|
7 |
+
"acc_stderr,none": 0.003551459735491157,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step106000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step106000",
|
71 |
+
"model_sha": "fb05d722aeeb0944d2f041429d68cd7a0f0aefb5",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761324.5612192,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143892.700193422,
|
94 |
+
"end_time": 1143972.705910608,
|
95 |
+
"total_evaluation_time_seconds": "80.00571718602441"
|
96 |
+
}
|
pythia-14m-seed0/step107000/EleutherAI__pythia-14m/results_2024-08-15T15-38-00.033086.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2030.1912500934511,
|
5 |
+
"perplexity_stderr,none": 89.98267591591406,
|
6 |
+
"acc,none": 0.05918882204541044,
|
7 |
+
"acc_stderr,none": 0.003287631671733364,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step107000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step107000",
|
71 |
+
"model_sha": "e51d508ae0521ac7f4d1f9879ae7a45133c4c71f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761412.6764872,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1143979.738471963,
|
94 |
+
"end_time": 1144060.273198181,
|
95 |
+
"total_evaluation_time_seconds": "80.5347262180876"
|
96 |
+
}
|
pythia-14m-seed0/step108000/EleutherAI__pythia-14m/results_2024-08-15T15-39-28.787439.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1942.191674587457,
|
5 |
+
"perplexity_stderr,none": 88.17936208188293,
|
6 |
+
"acc,none": 0.07219095672423831,
|
7 |
+
"acc_stderr,none": 0.003605643396603442,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step108000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step108000",
|
71 |
+
"model_sha": "e6e41d34d4cea7f1b7b7a3c789f10c07e9ea998a",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761500.0742729,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144067.406098665,
|
94 |
+
"end_time": 1144149.027880051,
|
95 |
+
"total_evaluation_time_seconds": "81.62178138596937"
|
96 |
+
}
|
pythia-14m-seed0/step109000/EleutherAI__pythia-14m/results_2024-08-15T15-40-55.857604.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1856.8399370102884,
|
5 |
+
"perplexity_stderr,none": 85.6448562808026,
|
6 |
+
"acc,none": 0.07413157384048127,
|
7 |
+
"acc_stderr,none": 0.0036499618074347898,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step109000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step109000",
|
71 |
+
"model_sha": "1406e20c840393345a50b6b8eaa261e2a8afa13f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761588.1894023,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144156.112510001,
|
94 |
+
"end_time": 1144236.097111593,
|
95 |
+
"total_evaluation_time_seconds": "79.98460159194656"
|
96 |
+
}
|
pythia-14m-seed0/step11000/EleutherAI__pythia-14m/results_2024-08-15T13-19-30.603325.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1336.2520161255652,
|
5 |
+
"perplexity_stderr,none": 60.98331616528402,
|
6 |
+
"acc,none": 0.10285270716087716,
|
7 |
+
"acc_stderr,none": 0.004232059957170525,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step11000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step11000",
|
71 |
+
"model_sha": "23b9f9586a7ce1aa3b85566361c5471f05e36f30",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723753103.9197896,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1135673.004151162,
|
94 |
+
"end_time": 1135750.843636572,
|
95 |
+
"total_evaluation_time_seconds": "77.83948540990241"
|
96 |
+
}
|
pythia-14m-seed0/step110000/EleutherAI__pythia-14m/results_2024-08-15T15-42-21.675993.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1655.1447995081642,
|
5 |
+
"perplexity_stderr,none": 76.78861282594242,
|
6 |
+
"acc,none": 0.08092373374733165,
|
7 |
+
"acc_stderr,none": 0.0037994943646444323,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step110000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step110000",
|
71 |
+
"model_sha": "4f102eb31047258ae88f77a33ad7a1d22107ad7f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761674.655817,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144243.183542134,
|
94 |
+
"end_time": 1144321.915977644,
|
95 |
+
"total_evaluation_time_seconds": "78.73243551002815"
|
96 |
+
}
|
pythia-14m-seed0/step111000/EleutherAI__pythia-14m/results_2024-08-15T15-43-46.617900.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1577.7555976392246,
|
5 |
+
"perplexity_stderr,none": 73.319450716533,
|
6 |
+
"acc,none": 0.08286435086357462,
|
7 |
+
"acc_stderr,none": 0.0038407207037192497,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step111000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step111000",
|
71 |
+
"model_sha": "99137989374bfe1ecb1a5e64b3d80ce774eec113",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761759.958622,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144328.95044279,
|
94 |
+
"end_time": 1144406.857867798,
|
95 |
+
"total_evaluation_time_seconds": "77.90742500801571"
|
96 |
+
}
|
pythia-14m-seed0/step112000/EleutherAI__pythia-14m/results_2024-08-15T15-45-12.514667.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1759.9239677979742,
|
5 |
+
"perplexity_stderr,none": 81.70703449195138,
|
6 |
+
"acc,none": 0.07490782068697846,
|
7 |
+
"acc_stderr,none": 0.0036674834158285856,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step112000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step112000",
|
71 |
+
"model_sha": "9774065c9f5a45ad7788ad185541011cbb0862d9",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761845.0600786,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144414.160817905,
|
94 |
+
"end_time": 1144492.75489787,
|
95 |
+
"total_evaluation_time_seconds": "78.59407996479422"
|
96 |
+
}
|
pythia-14m-seed0/step113000/EleutherAI__pythia-14m/results_2024-08-15T15-46-38.770492.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1588.236297320137,
|
5 |
+
"perplexity_stderr,none": 72.93322208929094,
|
6 |
+
"acc,none": 0.07490782068697846,
|
7 |
+
"acc_stderr,none": 0.0036674834158285856,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step113000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step113000",
|
71 |
+
"model_sha": "550e364271f369edb9d4a144425830fdedee75cd",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723761930.9352212,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144499.950961815,
|
94 |
+
"end_time": 1144579.010207957,
|
95 |
+
"total_evaluation_time_seconds": "79.05924614192918"
|
96 |
+
}
|
pythia-14m-seed0/step114000/EleutherAI__pythia-14m/results_2024-08-15T15-48-05.260016.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1702.065327436129,
|
5 |
+
"perplexity_stderr,none": 78.56178715398188,
|
6 |
+
"acc,none": 0.07568406753347565,
|
7 |
+
"acc_stderr,none": 0.0036848899715327707,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step114000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step114000",
|
71 |
+
"model_sha": "9c800cd9948d544aff8957bed479581938a7d59c",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762017.8288126,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144586.610589841,
|
94 |
+
"end_time": 1144665.499988281,
|
95 |
+
"total_evaluation_time_seconds": "78.88939844002016"
|
96 |
+
}
|
pythia-14m-seed0/step115000/EleutherAI__pythia-14m/results_2024-08-15T15-49-32.634482.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1653.5844188453527,
|
5 |
+
"perplexity_stderr,none": 76.53860729699839,
|
6 |
+
"acc,none": 0.07471375897535416,
|
7 |
+
"acc_stderr,none": 0.003663113864225891,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step115000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step115000",
|
71 |
+
"model_sha": "84565561177046e994cc5d236a3f526cb0a1494e",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762104.9888885,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144673.793967198,
|
94 |
+
"end_time": 1144752.874586123,
|
95 |
+
"total_evaluation_time_seconds": "79.08061892492697"
|
96 |
+
}
|
pythia-14m-seed0/step116000/EleutherAI__pythia-14m/results_2024-08-15T15-51-01.124488.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1670.6856831411806,
|
5 |
+
"perplexity_stderr,none": 76.90491802120467,
|
6 |
+
"acc,none": 0.07432563555210557,
|
7 |
+
"acc_stderr,none": 0.0036543530860415506,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step116000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step116000",
|
71 |
+
"model_sha": "90fd8b63997a429c96f4d13a9d5290cb9621af6e",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762193.411083,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144760.733506932,
|
94 |
+
"end_time": 1144841.363911434,
|
95 |
+
"total_evaluation_time_seconds": "80.6304045021534"
|
96 |
+
}
|
pythia-14m-seed0/step117000/EleutherAI__pythia-14m/results_2024-08-15T15-52-30.240629.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1754.3725242092196,
|
5 |
+
"perplexity_stderr,none": 81.01434898802123,
|
6 |
+
"acc,none": 0.08072967203570736,
|
7 |
+
"acc_stderr,none": 0.003795336508106496,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step117000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step117000",
|
71 |
+
"model_sha": "b205c539eccb75ec24cffe6c918b8eb5419e9e1f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762281.6944668,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144849.179324036,
|
94 |
+
"end_time": 1144930.480355115,
|
95 |
+
"total_evaluation_time_seconds": "81.30103107891046"
|
96 |
+
}
|
pythia-14m-seed0/step118000/EleutherAI__pythia-14m/results_2024-08-15T15-53-56.953504.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1758.961983552553,
|
5 |
+
"perplexity_stderr,none": 80.65986006671504,
|
6 |
+
"acc,none": 0.07587812924509994,
|
7 |
+
"acc_stderr,none": 0.0036892238245559935,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step118000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step118000",
|
71 |
+
"model_sha": "0cd8490766c1ec80c1f9aa970de646bd573192d1",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762369.5567913,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1144937.75132631,
|
94 |
+
"end_time": 1145017.193742402,
|
95 |
+
"total_evaluation_time_seconds": "79.44241609191522"
|
96 |
+
}
|
pythia-14m-seed0/step119000/EleutherAI__pythia-14m/results_2024-08-15T15-55-23.128925.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1654.0146622462084,
|
5 |
+
"perplexity_stderr,none": 76.2436881269689,
|
6 |
+
"acc,none": 0.07665437609159713,
|
7 |
+
"acc_stderr,none": 0.0037064888417067503,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step119000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step119000",
|
71 |
+
"model_sha": "06f016ef45fd4df8e4118bec1aaaa17ce539d8eb",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762455.409844,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145024.27587102,
|
94 |
+
"end_time": 1145103.369137769,
|
95 |
+
"total_evaluation_time_seconds": "79.09326674905606"
|
96 |
+
}
|
pythia-14m-seed0/step12000/EleutherAI__pythia-14m/results_2024-08-15T13-20-56.302395.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1345.9520903737475,
|
5 |
+
"perplexity_stderr,none": 62.175751078363405,
|
6 |
+
"acc,none": 0.09179118959829226,
|
7 |
+
"acc_stderr,none": 0.004022586828929574,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step12000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step12000",
|
71 |
+
"model_sha": "677c73e05b0353916ad35c85c4f4609dce0348af",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723753189.099923,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2200.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1135757.985463659,
|
94 |
+
"end_time": 1135836.540199971,
|
95 |
+
"total_evaluation_time_seconds": "78.55473631201312"
|
96 |
+
}
|
pythia-14m-seed0/step120000/EleutherAI__pythia-14m/results_2024-08-15T15-56-52.216443.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1885.6556307949313,
|
5 |
+
"perplexity_stderr,none": 84.00187419647823,
|
6 |
+
"acc,none": 0.06462254997089074,
|
7 |
+
"acc_stderr,none": 0.003425291731867342,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step120000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step120000",
|
71 |
+
"model_sha": "379be14e11d7e3ef68044bbbcc8890e192c013b9",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762541.51547,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145110.479581879,
|
94 |
+
"end_time": 1145192.456666261,
|
95 |
+
"total_evaluation_time_seconds": "81.97708438197151"
|
96 |
+
}
|
pythia-14m-seed0/step121000/EleutherAI__pythia-14m/results_2024-08-15T15-58-17.456909.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1760.6364312354708,
|
5 |
+
"perplexity_stderr,none": 81.10863398516484,
|
6 |
+
"acc,none": 0.07781874636134291,
|
7 |
+
"acc_stderr,none": 0.0037321778637123253,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step121000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step121000",
|
71 |
+
"model_sha": "b4462c932e24f3271ec4e0819a6af0af31105a9a",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762630.560148,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145199.569018539,
|
94 |
+
"end_time": 1145277.697085722,
|
95 |
+
"total_evaluation_time_seconds": "78.12806718284264"
|
96 |
+
}
|
pythia-14m-seed0/step122000/EleutherAI__pythia-14m/results_2024-08-15T15-59-44.315496.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1614.3906537485614,
|
5 |
+
"perplexity_stderr,none": 75.54730571478007,
|
6 |
+
"acc,none": 0.0836405977100718,
|
7 |
+
"acc_stderr,none": 0.0038570347957365973,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step122000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step122000",
|
71 |
+
"model_sha": "d05207ca4e2e9503facd5d5f03130fec1accc174",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762715.8275197,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145284.959185133,
|
94 |
+
"end_time": 1145364.550398592,
|
95 |
+
"total_evaluation_time_seconds": "79.59121345891617"
|
96 |
+
}
|
pythia-14m-seed0/step123000/EleutherAI__pythia-14m/results_2024-08-15T16-01-10.335351.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2038.1899485513527,
|
5 |
+
"perplexity_stderr,none": 91.65050326709839,
|
6 |
+
"acc,none": 0.066757228798758,
|
7 |
+
"acc_stderr,none": 0.0034774311506045566,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step123000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step123000",
|
71 |
+
"model_sha": "bf4ec8affb1219b69a78d3ea02f4300e65c2627e",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762802.8322046,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145371.838932145,
|
94 |
+
"end_time": 1145450.57434153,
|
95 |
+
"total_evaluation_time_seconds": "78.73540938505903"
|
96 |
+
}
|
pythia-14m-seed0/step124000/EleutherAI__pythia-14m/results_2024-08-15T16-02-35.938574.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1899.7256494472654,
|
5 |
+
"perplexity_stderr,none": 85.27743433714627,
|
6 |
+
"acc,none": 0.0671453522220066,
|
7 |
+
"acc_stderr,none": 0.00348680002818163,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step124000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step124000",
|
71 |
+
"model_sha": "e6fb05f7d7a0437fe0bed7c289eeec6034f3ec93",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762888.69861,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145457.648579555,
|
94 |
+
"end_time": 1145536.161768229,
|
95 |
+
"total_evaluation_time_seconds": "78.51318867388181"
|
96 |
+
}
|
pythia-14m-seed0/step125000/EleutherAI__pythia-14m/results_2024-08-15T16-04-02.754305.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1716.8749780802998,
|
5 |
+
"perplexity_stderr,none": 79.70990558710086,
|
6 |
+
"acc,none": 0.07607219095672424,
|
7 |
+
"acc_stderr,none": 0.0036935506133658606,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step125000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step125000",
|
71 |
+
"model_sha": "460b51a3b8394f0bbcc5f839d989f2e19a262a9f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723762975.4865017,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145543.265744231,
|
94 |
+
"end_time": 1145622.994148452,
|
95 |
+
"total_evaluation_time_seconds": "79.72840422112495"
|
96 |
+
}
|
pythia-14m-seed0/step126000/EleutherAI__pythia-14m/results_2024-08-15T16-05-29.292044.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1648.377668467864,
|
5 |
+
"perplexity_stderr,none": 76.19708183106972,
|
6 |
+
"acc,none": 0.07704249951484572,
|
7 |
+
"acc_stderr,none": 0.003715079456679115,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step126000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step126000",
|
71 |
+
"model_sha": "7bdc419ff72691c556674155c1a83bc6e88cd742",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763062.0582552,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145630.007922808,
|
94 |
+
"end_time": 1145709.532083368,
|
95 |
+
"total_evaluation_time_seconds": "79.52416055998765"
|
96 |
+
}
|
pythia-14m-seed0/step127000/EleutherAI__pythia-14m/results_2024-08-15T16-06-56.433830.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1746.9065161719782,
|
5 |
+
"perplexity_stderr,none": 80.67923580716102,
|
6 |
+
"acc,none": 0.07898311663108869,
|
7 |
+
"acc_stderr,none": 0.003757621238955833,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step127000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step127000",
|
71 |
+
"model_sha": "0ebce7cc0fad8445578b776dae9f0c2dd9017d50",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763149.0984063,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145716.698455785,
|
94 |
+
"end_time": 1145796.673910495,
|
95 |
+
"total_evaluation_time_seconds": "79.97545470995829"
|
96 |
+
}
|
pythia-14m-seed0/step128/EleutherAI__pythia-14m/results_2024-08-15T13-00-34.735877.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 2154450.549794474,
|
5 |
+
"perplexity_stderr,none": 190529.88059195384,
|
6 |
+
"acc,none": 0.0,
|
7 |
+
"acc_stderr,none": 0.0,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step128,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step128",
|
71 |
+
"model_sha": "712cec71d5646cb88f76ee92029c12132ccd754c",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723751967.8752835,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1134536.513561532,
|
94 |
+
"end_time": 1134614.975972693,
|
95 |
+
"total_evaluation_time_seconds": "78.46241116104648"
|
96 |
+
}
|
pythia-14m-seed0/step128000/EleutherAI__pythia-14m/results_2024-08-15T16-08-22.902858.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1695.180554031583,
|
5 |
+
"perplexity_stderr,none": 77.5396989490735,
|
6 |
+
"acc,none": 0.06908596933824956,
|
7 |
+
"acc_stderr,none": 0.003533147663272781,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step128000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step128000",
|
71 |
+
"model_sha": "698a9b44b5707c15193cb9462f02c1952bf7ef29",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763234.9688888,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145803.817017457,
|
94 |
+
"end_time": 1145883.143051203,
|
95 |
+
"total_evaluation_time_seconds": "79.32603374612518"
|
96 |
+
}
|
pythia-14m-seed0/step129000/EleutherAI__pythia-14m/results_2024-08-15T16-09-48.457182.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1748.2060562452823,
|
5 |
+
"perplexity_stderr,none": 80.70472934165429,
|
6 |
+
"acc,none": 0.06908596933824956,
|
7 |
+
"acc_stderr,none": 0.003533147663272781,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step129000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step129000",
|
71 |
+
"model_sha": "9e3f931ba445a319b202efae7ca3b6624b085ae4",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763321.257125,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145890.282894991,
|
94 |
+
"end_time": 1145968.695761718,
|
95 |
+
"total_evaluation_time_seconds": "78.41286672698334"
|
96 |
+
}
|
pythia-14m-seed0/step13000/EleutherAI__pythia-14m/results_2024-08-15T13-22-24.295538.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1245.8883648802528,
|
5 |
+
"perplexity_stderr,none": 57.69527819537371,
|
6 |
+
"acc,none": 0.10576363283524161,
|
7 |
+
"acc_stderr,none": 0.004284561852149411,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step13000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step13000",
|
71 |
+
"model_sha": "51725d80c402ddd0b1ef23cb322f0fff2b7ab057",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723753276.8532705,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1135845.439155096,
|
94 |
+
"end_time": 1135924.535229917,
|
95 |
+
"total_evaluation_time_seconds": "79.0960748209618"
|
96 |
+
}
|
pythia-14m-seed0/step130000/EleutherAI__pythia-14m/results_2024-08-15T16-11-14.614730.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1615.9022166310167,
|
5 |
+
"perplexity_stderr,none": 74.77044537396877,
|
6 |
+
"acc,none": 0.08111779545895595,
|
7 |
+
"acc_stderr,none": 0.0038036457543544435,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step130000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step130000",
|
71 |
+
"model_sha": "c5a452c0e21b2beb5583459f763449ed56caa5c2",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763407.04308,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1145975.991028769,
|
94 |
+
"end_time": 1146054.855413207,
|
95 |
+
"total_evaluation_time_seconds": "78.86438443814404"
|
96 |
+
}
|
pythia-14m-seed0/step131000/EleutherAI__pythia-14m/results_2024-08-15T16-12-40.692551.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1429.6528872004685,
|
5 |
+
"perplexity_stderr,none": 66.43871183549871,
|
6 |
+
"acc,none": 0.08888026392392781,
|
7 |
+
"acc_stderr,none": 0.003964628217610132,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step131000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step131000",
|
71 |
+
"model_sha": "4facff6a23be64c8b725719006ddd396b322144f",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763493.1769783,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146062.15420514,
|
94 |
+
"end_time": 1146140.932159504,
|
95 |
+
"total_evaluation_time_seconds": "78.77795436396264"
|
96 |
+
}
|
pythia-14m-seed0/step132000/EleutherAI__pythia-14m/results_2024-08-15T16-14-06.187317.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1744.4924353855185,
|
5 |
+
"perplexity_stderr,none": 81.16053763865396,
|
6 |
+
"acc,none": 0.07180283330098972,
|
7 |
+
"acc_stderr,none": 0.0035966897962539787,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step132000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step132000",
|
71 |
+
"model_sha": "d32e674e5bcaaabda91936dc3c8d6ebc2873fa18",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763579.0232463,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146148.026271967,
|
94 |
+
"end_time": 1146226.427460756,
|
95 |
+
"total_evaluation_time_seconds": "78.40118878893554"
|
96 |
+
}
|
pythia-14m-seed0/step133000/EleutherAI__pythia-14m/results_2024-08-15T16-15-33.241417.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1654.4108715342993,
|
5 |
+
"perplexity_stderr,none": 77.65858495677924,
|
6 |
+
"acc,none": 0.08228216572870173,
|
7 |
+
"acc_stderr,none": 0.0038284194650137684,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step133000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step133000",
|
71 |
+
"model_sha": "c0fe6645a6d36cae5975f2d5ba418a058ab1023e",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763665.9182217,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146233.538208775,
|
94 |
+
"end_time": 1146313.481529114,
|
95 |
+
"total_evaluation_time_seconds": "79.94332033907995"
|
96 |
+
}
|
pythia-14m-seed0/step134000/EleutherAI__pythia-14m/results_2024-08-15T16-17-00.743542.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1714.602030249706,
|
5 |
+
"perplexity_stderr,none": 78.98081491097976,
|
6 |
+
"acc,none": 0.07762468464971861,
|
7 |
+
"acc_stderr,none": 0.003727913555390672,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step134000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step134000",
|
71 |
+
"model_sha": "6aacc13c38d2aa88207071a8868e41c6e22cc82b",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763753.4132254,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146320.668391112,
|
94 |
+
"end_time": 1146400.983705253,
|
95 |
+
"total_evaluation_time_seconds": "80.31531414110214"
|
96 |
+
}
|
pythia-14m-seed0/step135000/EleutherAI__pythia-14m/results_2024-08-15T16-18-28.514020.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1645.9272622179124,
|
5 |
+
"perplexity_stderr,none": 76.43349328751616,
|
6 |
+
"acc,none": 0.07743062293809432,
|
7 |
+
"acc_stderr,none": 0.0037236424005269012,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step135000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step135000",
|
71 |
+
"model_sha": "95509b738e052a84f0d6741472fbec47268ada6d",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763840.8905025,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146408.240149021,
|
94 |
+
"end_time": 1146488.754164905,
|
95 |
+
"total_evaluation_time_seconds": "80.51401588413864"
|
96 |
+
}
|
pythia-14m-seed0/step136000/EleutherAI__pythia-14m/results_2024-08-15T16-19-54.399480.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1960.7084303102972,
|
5 |
+
"perplexity_stderr,none": 88.24537878239123,
|
6 |
+
"acc,none": 0.06442848825926645,
|
7 |
+
"acc_stderr,none": 0.003420499552612167,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step136000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step136000",
|
71 |
+
"model_sha": "2dd2b497d1fa1285e9288538bf03af70ea88c3f1",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723763927.2161357,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146495.818538224,
|
94 |
+
"end_time": 1146574.639468531,
|
95 |
+
"total_evaluation_time_seconds": "78.82093030679971"
|
96 |
+
}
|
pythia-14m-seed0/step137000/EleutherAI__pythia-14m/results_2024-08-15T16-21-20.940162.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1592.062887278126,
|
5 |
+
"perplexity_stderr,none": 73.0489453312968,
|
6 |
+
"acc,none": 0.07762468464971861,
|
7 |
+
"acc_stderr,none": 0.003727913555390672,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step137000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step137000",
|
71 |
+
"model_sha": "f9cd17305d5cfb4f3776c1ea96d9b0f00fe32ede",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723764012.7023847,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146581.770022683,
|
94 |
+
"end_time": 1146661.180396564,
|
95 |
+
"total_evaluation_time_seconds": "79.41037388099357"
|
96 |
+
}
|
pythia-14m-seed0/step138000/EleutherAI__pythia-14m/results_2024-08-15T16-22-46.483201.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1616.2517548162928,
|
5 |
+
"perplexity_stderr,none": 74.09170027590162,
|
6 |
+
"acc,none": 0.07490782068697846,
|
7 |
+
"acc_stderr,none": 0.0036674834158285856,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step138000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step138000",
|
71 |
+
"model_sha": "720f854dbe30d974b7b495c0b076e452aab854b9",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723764099.134047,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146668.278176658,
|
94 |
+
"end_time": 1146746.723473584,
|
95 |
+
"total_evaluation_time_seconds": "78.44529692595825"
|
96 |
+
}
|
pythia-14m-seed0/step139000/EleutherAI__pythia-14m/results_2024-08-15T16-24-12.061095.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1698.8500597374266,
|
5 |
+
"perplexity_stderr,none": 78.62022294671738,
|
6 |
+
"acc,none": 0.07549000582185135,
|
7 |
+
"acc_stderr,none": 0.00368054902934182,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step139000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step139000",
|
71 |
+
"model_sha": "7dde889802ebc8e9b3498bcc7d02ce199d9d4d76",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723764184.9015696,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146753.933885823,
|
94 |
+
"end_time": 1146832.300943617,
|
95 |
+
"total_evaluation_time_seconds": "78.36705779400654"
|
96 |
+
}
|
pythia-14m-seed0/step14000/EleutherAI__pythia-14m/results_2024-08-15T13-23-49.471021.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1334.3251049658206,
|
5 |
+
"perplexity_stderr,none": 62.74215047810956,
|
6 |
+
"acc,none": 0.10304676887250146,
|
7 |
+
"acc_stderr,none": 0.004235592410767947,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step14000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step14000",
|
71 |
+
"model_sha": "d1cd73f742163cf3d733776a53422349d945caa8",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723753362.6691036,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 2500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1135931.618820009,
|
94 |
+
"end_time": 1136009.711027219,
|
95 |
+
"total_evaluation_time_seconds": "78.09220720990561"
|
96 |
+
}
|
pythia-14m-seed0/step140000/EleutherAI__pythia-14m/results_2024-08-15T16-25-37.975454.json
ADDED
@@ -0,0 +1,96 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"results": {
|
3 |
+
"lambada_openai": {
|
4 |
+
"perplexity,none": 1635.2112969586547,
|
5 |
+
"perplexity_stderr,none": 75.2598585640202,
|
6 |
+
"acc,none": 0.07490782068697846,
|
7 |
+
"acc_stderr,none": 0.0036674834158285856,
|
8 |
+
"alias": "lambada_openai"
|
9 |
+
}
|
10 |
+
},
|
11 |
+
"group_subtasks": {
|
12 |
+
"lambada_openai": []
|
13 |
+
},
|
14 |
+
"configs": {
|
15 |
+
"lambada_openai": {
|
16 |
+
"task": "lambada_openai",
|
17 |
+
"group": [
|
18 |
+
"lambada"
|
19 |
+
],
|
20 |
+
"dataset_path": "EleutherAI/lambada_openai",
|
21 |
+
"dataset_name": "default",
|
22 |
+
"dataset_kwargs": {
|
23 |
+
"trust_remote_code": true
|
24 |
+
},
|
25 |
+
"test_split": "test",
|
26 |
+
"doc_to_text": "{{text.split(' ')[:-1]|join(' ')}}",
|
27 |
+
"doc_to_target": "{{' '+text.split(' ')[-1]}}",
|
28 |
+
"description": "",
|
29 |
+
"target_delimiter": " ",
|
30 |
+
"fewshot_delimiter": "\n\n",
|
31 |
+
"num_fewshot": 0,
|
32 |
+
"metric_list": [
|
33 |
+
{
|
34 |
+
"metric": "perplexity",
|
35 |
+
"aggregation": "perplexity",
|
36 |
+
"higher_is_better": false
|
37 |
+
},
|
38 |
+
{
|
39 |
+
"metric": "acc",
|
40 |
+
"aggregation": "mean",
|
41 |
+
"higher_is_better": true
|
42 |
+
}
|
43 |
+
],
|
44 |
+
"output_type": "loglikelihood",
|
45 |
+
"repeats": 1,
|
46 |
+
"should_decontaminate": true,
|
47 |
+
"doc_to_decontamination_query": "{{text}}",
|
48 |
+
"metadata": {
|
49 |
+
"version": 1.0
|
50 |
+
}
|
51 |
+
}
|
52 |
+
},
|
53 |
+
"versions": {
|
54 |
+
"lambada_openai": 1.0
|
55 |
+
},
|
56 |
+
"n-shot": {
|
57 |
+
"lambada_openai": 0
|
58 |
+
},
|
59 |
+
"n-samples": {
|
60 |
+
"lambada_openai": {
|
61 |
+
"original": 5153,
|
62 |
+
"effective": 5153
|
63 |
+
}
|
64 |
+
},
|
65 |
+
"config": {
|
66 |
+
"model": "hf",
|
67 |
+
"model_args": "pretrained=EleutherAI/pythia-14m,revision=step140000,",
|
68 |
+
"model_num_parameters": 14067712,
|
69 |
+
"model_dtype": "torch.float16",
|
70 |
+
"model_revision": "step140000",
|
71 |
+
"model_sha": "5bb0fea58b8b2712c579d791ff1c8bf2790cca10",
|
72 |
+
"batch_size": "128",
|
73 |
+
"batch_sizes": [],
|
74 |
+
"device": "cuda",
|
75 |
+
"use_cache": null,
|
76 |
+
"limit": null,
|
77 |
+
"bootstrap_iters": 100000,
|
78 |
+
"gen_kwargs": null,
|
79 |
+
"random_seed": 0,
|
80 |
+
"numpy_seed": 1234,
|
81 |
+
"torch_seed": 1234,
|
82 |
+
"fewshot_seed": 1234
|
83 |
+
},
|
84 |
+
"git_hash": "51a7ca9",
|
85 |
+
"date": 1723764270.5154667,
|
86 |
+
"pretty_env_info": "PyTorch version: 2.3.0+cu121\nIs debug build: False\nCUDA used to build PyTorch: 12.1\nROCM used to build PyTorch: N/A\n\nOS: CentOS Linux release 7.9.2009 (Core) (x86_64)\nGCC version: (GCC) 12.1.0\nClang version: Could not collect\nCMake version: Could not collect\nLibc version: glibc-2.17\n\nPython version: 3.12.1 (main, Jan 12 2024, 16:49:08) [GCC 12.1.0] (64-bit runtime)\nPython platform: Linux-3.10.0-1160.119.1.el7.tuxcare.els2.x86_64-x86_64-with-glibc2.17\nIs CUDA available: True\nCUDA runtime version: 12.4.99\nCUDA_MODULE_LOADING set to: LAZY\nGPU models and configuration: GPU 0: Tesla V100S-PCIE-32GB\nNvidia driver version: 550.90.07\ncuDNN version: Could not collect\nHIP runtime version: N/A\nMIOpen runtime version: N/A\nIs XNNPACK available: True\n\nCPU:\nArchitecture: x86_64\nCPU op-mode(s): 32-bit, 64-bit\nByte Order: Little Endian\nCPU(s): 32\nOn-line CPU(s) list: 0-31\nThread(s) per core: 1\nCore(s) per socket: 32\nSocket(s): 1\nNUMA node(s): 2\nVendor ID: AuthenticAMD\nCPU family: 23\nModel: 49\nModel name: AMD EPYC 7502P 32-Core Processor\nStepping: 0\nCPU MHz: 1500.000\nCPU max MHz: 2500.0000\nCPU min MHz: 1500.0000\nBogoMIPS: 5000.08\nVirtualization: AMD-V\nL1d cache: 32K\nL1i cache: 32K\nL2 cache: 512K\nL3 cache: 16384K\nNUMA node0 CPU(s): 0-15\nNUMA node1 CPU(s): 16-31\nFlags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush mmx fxsr sse sse2 ht syscall nx mmxext fxsr_opt pdpe1gb rdtscp lm constant_tsc art rep_good nopl nonstop_tsc extd_apicid aperfmperf eagerfpu pni pclmulqdq monitor ssse3 fma cx16 sse4_1 sse4_2 x2apic movbe popcnt aes xsave avx f16c rdrand lahf_lm cmp_legacy svm extapic cr8_legacy abm sse4a misalignsse 3dnowprefetch osvw ibs skinit wdt tce topoext perfctr_core perfctr_nb bpext perfctr_l2 cpb cat_l3 cdp_l3 hw_pstate sme ssbd ibrs ibpb stibp vmmcall fsgsbase bmi1 avx2 smep bmi2 cqm rdt_a rdseed adx smap clflushopt clwb sha_ni xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local clzero irperf xsaveerptr arat npt lbrv svm_lock nrip_save tsc_scale vmcb_clean flushbyasid decodeassists pausefilter pfthreshold avic v_vmsave_vmload vgif umip overflow_recov succor smca\n\nVersions of relevant libraries:\n[pip3] numpy==1.26.4\n[pip3] torch==2.3.0\n[conda] Could not collect",
|
87 |
+
"transformers_version": "4.40.2",
|
88 |
+
"upper_git_hash": null,
|
89 |
+
"task_hashes": {},
|
90 |
+
"model_source": "hf",
|
91 |
+
"model_name": "EleutherAI/pythia-14m",
|
92 |
+
"model_name_sanitized": "EleutherAI__pythia-14m",
|
93 |
+
"start_time": 1146839.438632164,
|
94 |
+
"end_time": 1146918.215870824,
|
95 |
+
"total_evaluation_time_seconds": "78.77723866002634"
|
96 |
+
}
|