|
--- |
|
base_model: |
|
- euclaise/Memphis-CoT-3B |
|
license: cc-by-4.0 |
|
datasets: |
|
- euclaise/TinyCoT |
|
- euclaise/mathoverflow-accepted |
|
- euclaise/reddit-instruct |
|
- euclaise/WritingPrompts_curated |
|
- sablo/oasst2_curated |
|
- euclaise/mathqa_programs |
|
- BEE-spoke-data/coedit-reworded-deduped |
|
- pszemraj/booksum-short |
|
library_name: transformers |
|
tags: |
|
- supertrainer2000 |
|
--- |
|
|
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64137e2150358a805203cbac/wEaKoLeJzidUdTWwQmA6k.png) |
|
|
|
Memphis-scribe 3B is a finetune of [Memphis-CoT 3B](https://huggingface.co/euclaise/Memphis-CoT-3B) on more creative data, which itself is a finetune of [StableLM 3B 4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t/). |
|
|
|
|
|
It is trained further on TinyCoT, but also on |
|
- 10000 comments from [reddit-instruct](https://hf.co/euclaise/reddit-instruct) |
|
- 15000 comments from [writingprompts-curated](https://hf.co/euclaise/writingprompts-curated) |
|
- [4000 cases of converting MathQA problems to Python snippets](https://hf.co/euclaise/mathqa_programs) |
|
- [3000 shorter rows of booksum (both chapter->summary and summary->chapter tasks)](https://huggingface.co/datasets/pszemraj/booksum-short) |
|
- 10000 examples from [mathoverflow-accepted](https://hf.co/euclaise/mathoverflow-accepted) comments with >10 upvotes |
|
- 1000 examples from [coedit-reworded-deduped](https://huggingface.co/datasets/BEE-spoke-data/coedit-reworded-deduped) |