File size: 1,652 Bytes
82ba686
91af3a9
 
0422044
 
 
82ba686
 
 
 
 
 
8c5c290
82ba686
 
8c5c290
82ba686
8c5c290
 
82ba686
 
 
 
 
 
 
de6eadb
 
 
 
1686826
 
26f3170
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f795301
26f3170
 
1686826
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
---
language:
- en
license: apache-2.0
task_categories:
- text-generation
dataset_info:
  features:
  - name: text
    dtype: string
  splits:
  - name: train
    num_bytes: 20912557
    num_examples: 9846
  - name: test
    num_bytes: 1108538
    num_examples: 518
  download_size: 10221850
  dataset_size: 22021095
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
  - split: test
    path: data/test-*
---

This dataset is [timdettmers/openassistant-guanaco](https://huggingface.co/datasets/timdettmers/openassistant-guanaco) converted to what I believe 
to be the Llama 2 prompt format (based on [this Reddit post](https://www.reddit.com/r/LocalLLaMA/comments/155po2p/get_llama_2_prompt_format_right/)).
It is otherwise unchanged.  

The format is like this:

```
<s>[INST] <<SYS>>
You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe.  Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.

If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.
<</SYS>>

{question} [/INST]
{response}
```

...where `{question}` is what the user asks first, and `{response}` is how the model is expected to respond.  Further
conversation turns are simply appended like this:

```
[INST]
{next_question}[/INST]
{next_response}
```

The license is the same as the original, Apache 2.0.