Update README.md
Browse files
README.md
CHANGED
@@ -17,9 +17,6 @@ The model architecture is Deberta V3 Base
|
|
17 |
Context length is 1024 tokens
|
18 |
# Training (details)
|
19 |
## Training data:
|
20 |
-
- 1 million Common Crawl samples, labeled using Google Cloud’s Natural Language API: https://cloud.google.com/natural-language/docs/classifying-text
|
21 |
-
- 500k Wikepedia articles, curated using Wikipedia-API: https://pypi.org/project/Wikipedia-API/
|
22 |
-
## Training steps:
|
23 |
The training set is 22828 Common Crawl text samples, labeled as "High", "Medium", "Low". Here are some examples:
|
24 |
1. Input:
|
25 |
```
|
|
|
17 |
Context length is 1024 tokens
|
18 |
# Training (details)
|
19 |
## Training data:
|
|
|
|
|
|
|
20 |
The training set is 22828 Common Crawl text samples, labeled as "High", "Medium", "Low". Here are some examples:
|
21 |
1. Input:
|
22 |
```
|