Update README.md
Browse files
README.md
CHANGED
@@ -10,7 +10,7 @@ license: mit
|
|
10 |
|
11 |
# SchorbGPT-Medium
|
12 |
|
13 |
-
This is a
|
14 |
|
15 |
## Model Details
|
16 |
|
@@ -75,7 +75,7 @@ The model achieves a word perplexity of 38.65 on WikiText, which is competitive
|
|
75 |
|
76 |
3. Linguistic Understanding:
|
77 |
- LAMBADA: 33.90% accuracy with perplexity of 36.21
|
78 |
-
- HellaSwag:
|
79 |
- Performance indicates basic linguistic and contextual understanding
|
80 |
- Typical range for non-fine-tuned models of this scale
|
81 |
|
|
|
10 |
|
11 |
# SchorbGPT-Medium
|
12 |
|
13 |
+
This is a medium sized language model trained on web data. The model uses the GPT-2 architecture and tokenizer.
|
14 |
|
15 |
## Model Details
|
16 |
|
|
|
75 |
|
76 |
3. Linguistic Understanding:
|
77 |
- LAMBADA: 33.90% accuracy with perplexity of 36.21
|
78 |
+
- HellaSwag: 31.26% (Random baseline: 25%)
|
79 |
- Performance indicates basic linguistic and contextual understanding
|
80 |
- Typical range for non-fine-tuned models of this scale
|
81 |
|