Datasets:
Tasks:
Text Generation
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
10K - 100K
License:
File size: 1,647 Bytes
82ba686 91af3a9 0422044 82ba686 8c5c290 82ba686 8c5c290 82ba686 8c5c290 82ba686 de6eadb 1686826 26f3170 1686826 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 |
---
language:
- en
license: apache-2.0
task_categories:
- text-generation
dataset_info:
features:
- name: text
dtype: string
splits:
- name: train
num_bytes: 20912557
num_examples: 9846
- name: test
num_bytes: 1108538
num_examples: 518
download_size: 10221850
dataset_size: 22021095
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
---
This dataset is [timdettmers/openassistant-guanaco](https://huggingface.co/datasets/timdettmers/openassistant-guanaco) converted to what I believe
to be the Llama 2 prompt format (based on [this Reddit post](https://www.reddit.com/r/LocalLLaMA/comments/155po2p/get_llama_2_prompt_format_right/)).
It is otherwise unchanged.
The format is like this:
```
<s>[INST] <<SYS>>
You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.
If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
<</SYS>>
{question} [/INST]
{response}
```
...where `{question}` is what the user asks first, and `{response}` is how the model is expected to respond. Further
conversation turns are simply appended like this:
```
[INST]
{next_question}[/INST]
{response}
```
The license is the same as the original, Apache 2.0. |