license: apache-2.0
language:
- en
tags:
- sft
pipeline_tag: text-generation
widget:
- text: >-
<|prompter|>What is a meme, and what's the history behind this
word?<|endoftext|><|assistant|>
- text: <|prompter|>What's the Earth total population<|endoftext|><|assistant|>
- text: >-
<|prompter|>Write a story about future of AI
development<|endoftext|><|assistant|>
Open-Assistant SFT-1 12B Model
This is the first iteration English supervised-fine-tuning (SFT) model of the Open-Assistant project. It is based on a Pythia 12B that was fine-tuned on ~22k human demonstrations of assistant conversations collected through the https://open-assistant.io/ human feedback web app before March 7, 2023.
Model Details
- Developed by: Open-Assistant Contributors
- Model type: Transformer-based Language Model
- Language: English
- Finetuned from: EleutherAI / pythia-12b-deduped
- Code: Open-Assistant/model/model_training
- Demo: Continuations for 250 random prompts (sampling code).
- License: Apache 2.0
- Contact: Open-Assistant Discord
Prompting
Two special tokens are used to mark the beginning of user and assistant turns:
<|prompter|>
and <|assistant|>
. Each turn ends with a <|endoftext|>
token.
Input prompt example:
<|prompter|>What is a meme, and what's the history behind this word?<|endoftext|><|assistant|>
The input ends with the <|assistant|>
token to signal that the model should
start generating the assistant reply.
Example of generated continuation (typical sampling 0.2):
A visual joke often accompanied by text that spreads rapidly from individual
to individual. The exact origin of the term "meme" is unclear, but it was
first used in the 1990s by Richard Dawkins in his book "The Selfish Gene."
He defined a meme as an idea, behavior, or style that is transmitted from
one person to another within a culture.
Dawkins' use of the term caught on among scholars and journalists, who began
to apply the concept to other types of cultural items such as advertisements,
fashion trends, and jokes. Over time, the term "meme" has come to be used
more generally to describe any social behavior that can be repeated and
altered by individuals. Today, the term is widely recognized and applied in
fields ranging from anthropology to psychology to marketing.<|endoftext|>
Limitations
See limitations of Pythia 12B base model here.
The model is known to fail horribly at answering math and coding questions.
Beware of hallucinations: Outputs are often factually wrong or misleading. Replies might look convincing (at first glance) while containing completely made up false statements.
This model is usable only for English conversations.
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 35.84 |
ARC (25-shot) | 46.42 |
HellaSwag (10-shot) | 70.0 |
MMLU (5-shot) | 26.19 |
TruthfulQA (0-shot) | 39.19 |
Winogrande (5-shot) | 62.19 |
GSM8K (5-shot) | 0.61 |
DROP (3-shot) | 6.3 |