|
--- |
|
license: cc-by-4.0 |
|
datasets: |
|
- HuggingFaceH4/no_robots |
|
- Ba2han/Reddit-instruct-curated_rated-1.2k |
|
- Ba2han/databricks-dolly_rated |
|
- Open-Orca/OpenOrca |
|
--- |
|
|
|
 |
|
The training dataset consists of 2k (longest) examples from no_robots, reddit_instruct, dolly, OpenOrca plus two other personal datasets. |
|
|
|
Please use with ChatML and the default system message or enter your own. It was trained with various system messages, the one in the config being the default one. |
|
|
|
The model is: |
|
- Very good at generating long and coherent text. |
|
|
|
- Creative due to data from Reddit ELI5 and a few other sources. |
|
|
|
- Better at handling longer input. |
|
|
|
- Not great with short text both in input and generation. |
|
|
|
The aim is to see how the **"Long is More for Alignment"** paper holds. This is basically a combination of LIMA + LMA. There should be no benchmark contamination as far as I am aware of. Around 70% of the data is from the mentioned datasets. I am happy with how it turned out. |
|
|
|
|
|
 |
|
|
|
|