license: apache-2.0 | |
datasets: | |
- shibing624/sharegpt_gpt4 | |
language: | |
- en | |
- zh | |
- fr | |
- es | |
# Model Card | |
Pythia-70m-deduped finetuned on a cleaned version of ShareGPT data. | |
<br>The cleaned dataset is obtained by removing duplicates and paraphrases from the original corpus. The final training size is of 3494 instances. | |