Update README.md
Browse files
README.md
CHANGED
@@ -13,22 +13,8 @@ tags:
|
|
13 |
- code
|
14 |
|
15 |
---
|
16 |
-
license: mit
|
17 |
-
datasets:
|
18 |
-
- Canstralian/ShellCommands
|
19 |
-
- Canstralian/CyberExploitDB
|
20 |
-
language:
|
21 |
-
- en
|
22 |
-
base_model:
|
23 |
-
- WhiteRabbitNeo/WhiteRabbitNeo-13B-v1
|
24 |
-
- replit/replit-code-v1_5-3b
|
25 |
-
library_name: transformers
|
26 |
-
tags:
|
27 |
-
- code
|
28 |
-
---
|
29 |
-
# Model Card for Model ID
|
30 |
|
31 |
-
|
32 |
|
33 |
This model card aims to document the capabilities, performance, and intended usage of models fine-tuned for cybersecurity tasks, including shell command parsing and cyber exploit detection. It is based on the underlying models WhiteRabbitNeo-13B-v1 and replit-code-v1_5-3b, fine-tuned on datasets related to shell commands and exploit databases.
|
34 |
|
@@ -38,19 +24,11 @@ This model card aims to document the capabilities, performance, and intended usa
|
|
38 |
|
39 |
This model is a fine-tuned version of large-scale language models optimized for tasks such as parsing shell commands and analyzing cybersecurity exploits. The training leverages datasets such as Canstralian/ShellCommands and Canstralian/CyberExploitDB to provide domain-specific knowledge.
|
40 |
|
41 |
-
- **Developed by:** Canstralian
|
42 |
-
- **Funded by [optional]:** [More Information Needed]
|
43 |
-
- **Shared by [optional]:** [More Information Needed]
|
44 |
- **Model type:** Transformer-based Language Model for cybersecurity applications
|
45 |
- **Language(s) (NLP):** English (en)
|
46 |
- **License:** MIT
|
47 |
-
- **Finetuned from model
|
48 |
-
|
49 |
-
### Model Sources [optional]
|
50 |
-
|
51 |
-
- **Repository:** [Add model repository URL here]
|
52 |
-
- **Paper [optional]:** [Link to relevant research paper]
|
53 |
-
- **Demo [optional]:** [Link to model demo or interface]
|
54 |
|
55 |
## Uses
|
56 |
|
@@ -61,7 +39,7 @@ The model is intended to be used directly for tasks like:
|
|
61 |
- Analyzing and classifying cybersecurity exploit patterns
|
62 |
- Assisting with code generation and debugging in a cybersecurity context
|
63 |
|
64 |
-
### Downstream Use
|
65 |
|
66 |
When fine-tuned further, the model can be applied to:
|
67 |
- Automated incident response systems
|
@@ -112,7 +90,7 @@ Further details on the preprocessing of these datasets can be found in their res
|
|
112 |
|
113 |
### Training Procedure
|
114 |
|
115 |
-
#### Preprocessing
|
116 |
|
117 |
The data was preprocessed to remove any sensitive or personally identifiable information. Text normalization and tokenization were applied to ensure consistency across the datasets.
|
118 |
|
@@ -120,12 +98,6 @@ The data was preprocessed to remove any sensitive or personally identifiable inf
|
|
120 |
|
121 |
- **Training regime:** fp16 mixed precision
|
122 |
|
123 |
-
#### Speeds, Sizes, Times [optional]
|
124 |
-
|
125 |
-
- **Training time:** [More Information Needed]
|
126 |
-
- **Model size:** [More Information Needed]
|
127 |
-
- **Dataset size:** [More Information Needed]
|
128 |
-
|
129 |
## Evaluation
|
130 |
|
131 |
### Testing Data, Factors & Metrics
|
@@ -170,36 +142,6 @@ Carbon emissions can be estimated using the [Machine Learning Impact calculator]
|
|
170 |
- **Compute Region:** [More Information Needed]
|
171 |
- **Carbon Emitted:** [More Information Needed]
|
172 |
|
173 |
-
## Technical Specifications [optional]
|
174 |
-
|
175 |
-
### Model Architecture and Objective
|
176 |
-
|
177 |
-
This model utilizes transformer-based architecture with a focus on optimizing the understanding of shell commands and cybersecurity exploits.
|
178 |
-
|
179 |
-
### Compute Infrastructure
|
180 |
-
|
181 |
-
The model was trained on [specify hardware and infrastructure used, e.g., GPUs, TPUs, cloud services].
|
182 |
-
|
183 |
-
#### Hardware
|
184 |
-
|
185 |
-
- [More Information Needed]
|
186 |
-
|
187 |
-
#### Software
|
188 |
-
|
189 |
-
- [More Information Needed]
|
190 |
-
|
191 |
-
## Citation [optional]
|
192 |
-
|
193 |
-
If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section.
|
194 |
-
|
195 |
-
**BibTeX:**
|
196 |
-
|
197 |
-
[More Information Needed]
|
198 |
-
|
199 |
-
**APA:**
|
200 |
-
|
201 |
-
[More Information Needed]
|
202 |
-
|
203 |
## Glossary [optional]
|
204 |
|
205 |
Terms like "shell command", "exploit", and "cybersecurity" may be used frequently in this model's context. Further definitions will help readers understand model performance.
|
|
|
13 |
- code
|
14 |
|
15 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
|
17 |
+
# Model Card for Model ID
|
18 |
|
19 |
This model card aims to document the capabilities, performance, and intended usage of models fine-tuned for cybersecurity tasks, including shell command parsing and cyber exploit detection. It is based on the underlying models WhiteRabbitNeo-13B-v1 and replit-code-v1_5-3b, fine-tuned on datasets related to shell commands and exploit databases.
|
20 |
|
|
|
24 |
|
25 |
This model is a fine-tuned version of large-scale language models optimized for tasks such as parsing shell commands and analyzing cybersecurity exploits. The training leverages datasets such as Canstralian/ShellCommands and Canstralian/CyberExploitDB to provide domain-specific knowledge.
|
26 |
|
27 |
+
- **Developed by:** Canstralian
|
|
|
|
|
28 |
- **Model type:** Transformer-based Language Model for cybersecurity applications
|
29 |
- **Language(s) (NLP):** English (en)
|
30 |
- **License:** MIT
|
31 |
+
- **Finetuned from model:** WhiteRabbitNeo/WhiteRabbitNeo-13B-v1, replit/replit-code-v1_5-3b
|
|
|
|
|
|
|
|
|
|
|
|
|
32 |
|
33 |
## Uses
|
34 |
|
|
|
39 |
- Analyzing and classifying cybersecurity exploit patterns
|
40 |
- Assisting with code generation and debugging in a cybersecurity context
|
41 |
|
42 |
+
### Downstream Use
|
43 |
|
44 |
When fine-tuned further, the model can be applied to:
|
45 |
- Automated incident response systems
|
|
|
90 |
|
91 |
### Training Procedure
|
92 |
|
93 |
+
#### Preprocessing
|
94 |
|
95 |
The data was preprocessed to remove any sensitive or personally identifiable information. Text normalization and tokenization were applied to ensure consistency across the datasets.
|
96 |
|
|
|
98 |
|
99 |
- **Training regime:** fp16 mixed precision
|
100 |
|
|
|
|
|
|
|
|
|
|
|
|
|
101 |
## Evaluation
|
102 |
|
103 |
### Testing Data, Factors & Metrics
|
|
|
142 |
- **Compute Region:** [More Information Needed]
|
143 |
- **Carbon Emitted:** [More Information Needed]
|
144 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
145 |
## Glossary [optional]
|
146 |
|
147 |
Terms like "shell command", "exploit", and "cybersecurity" may be used frequently in this model's context. Further definitions will help readers understand model performance.
|