secbench commited on
Commit
4d343a1
·
verified ·
1 Parent(s): fe0d3d7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -35,7 +35,7 @@ SecBench Technical Paper: *[link](https://arxiv.org/abs/2412.20787)*.
35
 
36
 
37
 
38
- ## SecBench Design
39
 
40
 
41
  The following figure shows the overview of the SecBench design: it is a comprehensive benchmarking dataset aiming to benchmark LLM's capability in cybersecurity from *Multi-Level*, *Multi-Language*, *Multi-Form*, *Multi-Domain*.
@@ -55,7 +55,7 @@ The following figure shows the overview of the SecBench design: it is a comprehe
55
 
56
  - **Multi-Domain** : The questions in SecBench consist of 9 different domains, including **D1. Security Management**, **D2. Data Security**, **D3. Network and Infrastructure Security**, **D4. Security Standards and Regulations**, **D5. Application Security**, **D6. Identity and Access Control**, **D7. Fundamental Software and Hardware and Technology**, **D8. Endpoint and Host Security**, **D9. Cloud Security**. Particularly, the above domains were devised from several rounds of brainstorming and revision, which were expected to cover most (if not all) related sub-domains in cybersecurity. Note that we do not expect these domains to be \emph{orthogonal}, and it is possible that one question can be reasonably labeled into different domains. In our dataset, one question is assigned only one most-related domain label from D1 to D9.
57
 
58
- ## Data Example
59
 
60
 
61
  ### MCQ Example
@@ -76,7 +76,7 @@ Following is one SAQ example, labeled in the domain of *Data Security* and the l
76
 
77
 
78
 
79
- ## Benchmarking
80
 
81
  Based on SecBench, we conducted extensive benchmarking on 16 SOTA LLMs, including the GPT series and competitive open-source ones.
82
 
@@ -91,7 +91,7 @@ Based on SecBench, we conducted extensive benchmarking on 16 SOTA LLMs, includin
91
 
92
 
93
 
94
- ## Released Data
95
 
96
  We release a total of 3,000 questions from SecBench (under the [data] folder), including:
97
 
 
35
 
36
 
37
 
38
+ ## 1, SecBench Design
39
 
40
 
41
  The following figure shows the overview of the SecBench design: it is a comprehensive benchmarking dataset aiming to benchmark LLM's capability in cybersecurity from *Multi-Level*, *Multi-Language*, *Multi-Form*, *Multi-Domain*.
 
55
 
56
  - **Multi-Domain** : The questions in SecBench consist of 9 different domains, including **D1. Security Management**, **D2. Data Security**, **D3. Network and Infrastructure Security**, **D4. Security Standards and Regulations**, **D5. Application Security**, **D6. Identity and Access Control**, **D7. Fundamental Software and Hardware and Technology**, **D8. Endpoint and Host Security**, **D9. Cloud Security**. Particularly, the above domains were devised from several rounds of brainstorming and revision, which were expected to cover most (if not all) related sub-domains in cybersecurity. Note that we do not expect these domains to be \emph{orthogonal}, and it is possible that one question can be reasonably labeled into different domains. In our dataset, one question is assigned only one most-related domain label from D1 to D9.
57
 
58
+ ## 2. Data Example
59
 
60
 
61
  ### MCQ Example
 
76
 
77
 
78
 
79
+ ## 3. Benchmarking
80
 
81
  Based on SecBench, we conducted extensive benchmarking on 16 SOTA LLMs, including the GPT series and competitive open-source ones.
82
 
 
91
 
92
 
93
 
94
+ ## 4. Released Data
95
 
96
  We release a total of 3,000 questions from SecBench (under the [data] folder), including:
97