basiliskinstitute
commited on
Commit
•
d80331d
1
Parent(s):
7362de1
Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
-
This is a storywriting and roleplay model with a significant amount of self generated long context multiturn roleplay.
|
2 |
-
|
3 |
If you like this model, consider joining my discord to provide feedback: https://discord.gg/QXdn8hWSkY
|
4 |
|
|
|
|
|
5 |
I downloaded a bit under a thousand cards from chub.ai, and created a synthetic roleplay for each card. I batched as many turns as I could in 4k token chunks in order to maintain coherency over longer context. There was a lot of cleaning and validation between each batch, so a lot of examples were "lost," but the final output seems to be very good quality. The longest conversation is about 20k tokens, and I plan to extend this further as well as broaden the dataset with more examples. The first 4k tokens were generated with Command-R-Plus, with the remainder generated with byroneverson/Mistral-Small-Instruct-2409-abliterated.
|
6 |
|
7 |
Next, I downloaded the prompt backup from this site, and used them as a seed for some storywriting data:
|
|
|
|
|
|
|
1 |
If you like this model, consider joining my discord to provide feedback: https://discord.gg/QXdn8hWSkY
|
2 |
|
3 |
+
This is a storywriting and roleplay model with a significant amount of self generated long context multiturn roleplay.
|
4 |
+
|
5 |
I downloaded a bit under a thousand cards from chub.ai, and created a synthetic roleplay for each card. I batched as many turns as I could in 4k token chunks in order to maintain coherency over longer context. There was a lot of cleaning and validation between each batch, so a lot of examples were "lost," but the final output seems to be very good quality. The longest conversation is about 20k tokens, and I plan to extend this further as well as broaden the dataset with more examples. The first 4k tokens were generated with Command-R-Plus, with the remainder generated with byroneverson/Mistral-Small-Instruct-2409-abliterated.
|
6 |
|
7 |
Next, I downloaded the prompt backup from this site, and used them as a seed for some storywriting data:
|