Transformers
English
code
Canstralian commited on
Commit
89f6140
·
verified ·
1 Parent(s): 45b9a79

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -63
README.md CHANGED
@@ -13,22 +13,8 @@ tags:
13
  - code
14
 
15
  ---
16
- license: mit
17
- datasets:
18
- - Canstralian/ShellCommands
19
- - Canstralian/CyberExploitDB
20
- language:
21
- - en
22
- base_model:
23
- - WhiteRabbitNeo/WhiteRabbitNeo-13B-v1
24
- - replit/replit-code-v1_5-3b
25
- library_name: transformers
26
- tags:
27
- - code
28
- ---
29
- # Model Card for Model ID
30
 
31
- <!-- Provide a quick summary of what the model is/does. -->
32
 
33
  This model card aims to document the capabilities, performance, and intended usage of models fine-tuned for cybersecurity tasks, including shell command parsing and cyber exploit detection. It is based on the underlying models WhiteRabbitNeo-13B-v1 and replit-code-v1_5-3b, fine-tuned on datasets related to shell commands and exploit databases.
34
 
@@ -38,19 +24,11 @@ This model card aims to document the capabilities, performance, and intended usa
38
 
39
  This model is a fine-tuned version of large-scale language models optimized for tasks such as parsing shell commands and analyzing cybersecurity exploits. The training leverages datasets such as Canstralian/ShellCommands and Canstralian/CyberExploitDB to provide domain-specific knowledge.
40
 
41
- - **Developed by:** Canstralian (more details needed)
42
- - **Funded by [optional]:** [More Information Needed]
43
- - **Shared by [optional]:** [More Information Needed]
44
  - **Model type:** Transformer-based Language Model for cybersecurity applications
45
  - **Language(s) (NLP):** English (en)
46
  - **License:** MIT
47
- - **Finetuned from model [optional]:** WhiteRabbitNeo/WhiteRabbitNeo-13B-v1, replit/replit-code-v1_5-3b
48
-
49
- ### Model Sources [optional]
50
-
51
- - **Repository:** [Add model repository URL here]
52
- - **Paper [optional]:** [Link to relevant research paper]
53
- - **Demo [optional]:** [Link to model demo or interface]
54
 
55
  ## Uses
56
 
@@ -61,7 +39,7 @@ The model is intended to be used directly for tasks like:
61
  - Analyzing and classifying cybersecurity exploit patterns
62
  - Assisting with code generation and debugging in a cybersecurity context
63
 
64
- ### Downstream Use [optional]
65
 
66
  When fine-tuned further, the model can be applied to:
67
  - Automated incident response systems
@@ -112,7 +90,7 @@ Further details on the preprocessing of these datasets can be found in their res
112
 
113
  ### Training Procedure
114
 
115
- #### Preprocessing [optional]
116
 
117
  The data was preprocessed to remove any sensitive or personally identifiable information. Text normalization and tokenization were applied to ensure consistency across the datasets.
118
 
@@ -120,12 +98,6 @@ The data was preprocessed to remove any sensitive or personally identifiable inf
120
 
121
  - **Training regime:** fp16 mixed precision
122
 
123
- #### Speeds, Sizes, Times [optional]
124
-
125
- - **Training time:** [More Information Needed]
126
- - **Model size:** [More Information Needed]
127
- - **Dataset size:** [More Information Needed]
128
-
129
  ## Evaluation
130
 
131
  ### Testing Data, Factors & Metrics
@@ -170,36 +142,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
170
  - **Compute Region:** [More Information Needed]
171
  - **Carbon Emitted:** [More Information Needed]
172
 
173
- ## Technical Specifications [optional]
174
-
175
- ### Model Architecture and Objective
176
-
177
- This model utilizes transformer-based architecture with a focus on optimizing the understanding of shell commands and cybersecurity exploits.
178
-
179
- ### Compute Infrastructure
180
-
181
- The model was trained on [specify hardware and infrastructure used, e.g., GPUs, TPUs, cloud services].
182
-
183
- #### Hardware
184
-
185
- - [More Information Needed]
186
-
187
- #### Software
188
-
189
- - [More Information Needed]
190
-
191
- ## Citation [optional]
192
-
193
- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section.
194
-
195
- **BibTeX:**
196
-
197
- [More Information Needed]
198
-
199
- **APA:**
200
-
201
- [More Information Needed]
202
-
203
  ## Glossary [optional]
204
 
205
  Terms like "shell command", "exploit", and "cybersecurity" may be used frequently in this model's context. Further definitions will help readers understand model performance.
 
13
  - code
14
 
15
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
+ # Model Card for Model ID
18
 
19
  This model card aims to document the capabilities, performance, and intended usage of models fine-tuned for cybersecurity tasks, including shell command parsing and cyber exploit detection. It is based on the underlying models WhiteRabbitNeo-13B-v1 and replit-code-v1_5-3b, fine-tuned on datasets related to shell commands and exploit databases.
20
 
 
24
 
25
  This model is a fine-tuned version of large-scale language models optimized for tasks such as parsing shell commands and analyzing cybersecurity exploits. The training leverages datasets such as Canstralian/ShellCommands and Canstralian/CyberExploitDB to provide domain-specific knowledge.
26
 
27
+ - **Developed by:** Canstralian
 
 
28
  - **Model type:** Transformer-based Language Model for cybersecurity applications
29
  - **Language(s) (NLP):** English (en)
30
  - **License:** MIT
31
+ - **Finetuned from model:** WhiteRabbitNeo/WhiteRabbitNeo-13B-v1, replit/replit-code-v1_5-3b
 
 
 
 
 
 
32
 
33
  ## Uses
34
 
 
39
  - Analyzing and classifying cybersecurity exploit patterns
40
  - Assisting with code generation and debugging in a cybersecurity context
41
 
42
+ ### Downstream Use
43
 
44
  When fine-tuned further, the model can be applied to:
45
  - Automated incident response systems
 
90
 
91
  ### Training Procedure
92
 
93
+ #### Preprocessing
94
 
95
  The data was preprocessed to remove any sensitive or personally identifiable information. Text normalization and tokenization were applied to ensure consistency across the datasets.
96
 
 
98
 
99
  - **Training regime:** fp16 mixed precision
100
 
 
 
 
 
 
 
101
  ## Evaluation
102
 
103
  ### Testing Data, Factors & Metrics
 
142
  - **Compute Region:** [More Information Needed]
143
  - **Carbon Emitted:** [More Information Needed]
144
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
145
  ## Glossary [optional]
146
 
147
  Terms like "shell command", "exploit", and "cybersecurity" may be used frequently in this model's context. Further definitions will help readers understand model performance.