A roleplay-based model intended for multiple AI characters / group based roleplaying sessions.

This is an experimental model, trained entirely off of purely human data, none from LLMs or AI Models. Data taken is from roleplaying forum scrapes and more, etc.

Built off of Llama-3-Instruct.

This is a Test / Alpha model. A proof of concept.

Make sure to adapt your cards to a group-chat friendly style?


Notes:

- Don't expect this to beat Stheno or other mature models. It won't.
- Works best when in group chat scenarios, with properly defined cards, I think it is a successful test?
- A very small dataset of varying quality (human data) was used. Does not work well outside of its specified scenario.

Training Details:

- Uses L3-Instruct Format.
- One designated character per entry is the Human-turn, while all other characters are different, assigned GPT-turns, not in order to simulate real group chats.
- Per entry, it varies from 2-5 Unique Characters usually.
- There is roughly only ~3K sample entries.
- May not be the smartest due to all samples being roleplay / conversational data.

Potential Issues:

- 1 on 1 RP Performance might be Affected as focus is solely on group chats.
- The names may be multiple tokens instead of one token as they replace User / Assistant -> May Affect Output Quality -> Another Idea is in the works.
- Dataset Quality? While it is filtered, a few times... there's still the occasional low quality in there. I have not gone through a manual pass, this is a proof of concept.

SillyTavern Settings:

Llama-3-Instruct-With-Names -->>>> Remove the Square Brackets in `[{{name}}]` or `[{{char}}]` or `[{{user}}]` within the instruction template to match the format used for training.

Example Dataset Entry:

    {
        "token_length": x,
        "Unique_chars": 3,
        "conversations": [
            {
                "from": "system",
                "value": "text"
            },
            {
                "from": "human-chat",
                "name": "User-1",
                "value": "text"
            },
            {
                "from": "gpt-chat",
                "name": "User-2",
                "value": "text"
            },
            {
                "from": "human-chat",
                "name": "User-1",
                "value": "text"
            },
            {
                "from": "gpt-chat",
                "name": "User-3",
                "value": "text"
            },
            {
                "from": "human-chat",
                "name": "User-1",
                "value": "text"
            },
            {
                "from": "gpt-chat",
                "name": "User-2",
                "value": "text"
            },
            {
                "from": "gpt-chat",
                "name": "User-3",
                "value": "text"
            }
        ]
    },
Downloads last month
19
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for Sao10K/L3-8B-Chara-v1-Alpha

Finetunes
5 models
Quantizations
6 models