Update README.md
Browse files
README.md
CHANGED
@@ -1,5 +1,48 @@
|
|
1 |
-
---
|
2 |
-
license: other
|
3 |
-
license_name: mrl
|
4 |
-
license_link: https://mistral.ai/licenses/MRL-0.1.md
|
5 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: other
|
3 |
+
license_name: mrl
|
4 |
+
license_link: https://mistral.ai/licenses/MRL-0.1.md
|
5 |
+
tags:
|
6 |
+
- not-for-all-audiences
|
7 |
+
- nsfw
|
8 |
+
language:
|
9 |
+
- ja
|
10 |
+
---
|
11 |
+
|
12 |
+
# Kudryavka 8B Alpha
|
13 |
+
|
14 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/6513cf1397fb08378dea8f3a/0R3JH9xBNcuYpn8Apy6Uo.jpeg" alt="drawing" style="width:512px;"/>
|
15 |
+
|
16 |
+
|
17 |
+
|
18 |
+
## Base Model
|
19 |
+
This model is built upon [Ministral-8B-Instruct-2410](https://huggingface.co/mistralai/Ministral-8B-Instruct-2410)
|
20 |
+
|
21 |
+
### NOTICE
|
22 |
+
|
23 |
+
Licensed by Mistral AI under the Mistral AI Research License.
|
24 |
+
By using this model, you agree to be bound by the terms of the Mistral AI Research License.
|
25 |
+
|
26 |
+
## Description
|
27 |
+
|
28 |
+
First and foremost, this model is under development and currently falls far short of the Berghof series.
|
29 |
+
|
30 |
+
I chose the 8B model simply because in my environment, it's difficult to load models larger than 9B (without quantization), and I wanted to select the largest size that I could handle.
|
31 |
+
|
32 |
+
Gemma2 9B was also an option, but I decided to pass on it after testing.
|
33 |
+
|
34 |
+
This model has the following characteristics:
|
35 |
+
|
36 |
+
- It can write relatively consistent continuations if given a sufficiently long prompt (1000~ tokens) beforehand;
|
37 |
+
- It tends to be roundabout and verbose, sometimes providing excessive explanations in the narrative text, often resulting in a slower-paced development
|
38 |
+
- It occasionally generates NSFW content in unexpected ways
|
39 |
+
- Its output undoubtedly tends to be repetitive, especially when the initial prompt is sparse
|
40 |
+
- If often fails to insert linebreaks between dialogues and monologues. This is expected to be fixed in later release of the series(hopefully)
|
41 |
+
|
42 |
+
As a rule of thumb, I think it's suitable for situations where you have a relatively complete piece of writing and want help continuing it, or when writing collaboratively while manually correcting inconsistencies here and there.
|
43 |
+
And for those who don't know, repetitiveness can be greatly reduced by using the DRY multiplier.
|
44 |
+
|
45 |
+
## Intended Use
|
46 |
+
Primarily designed for novel generation. **Not** optimized for:
|
47 |
+
- Role-playing (RP) scenarios
|
48 |
+
- Instruction-based responses
|