--- base_model: - euclaise/Memphis-CoT-3B license: cc-by-4.0 datasets: - euclaise/TinyCoT - euclaise/mathoverflow-accepted - euclaise/reddit-instruct - euclaise/WritingPrompts_curated - sablo/oasst2_curated - euclaise/mathqa_programs - BEE-spoke-data/coedit-reworded-deduped - pszemraj/booksum-short library_name: transformers tags: - supertrainer2000 --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64137e2150358a805203cbac/wEaKoLeJzidUdTWwQmA6k.png) Memphis-scribe 3B is a finetune of [Memphis-CoT 3B](https://huggingface.co/euclaise/Memphis-CoT-3B) on more creative data, which itself is a finetune of [StableLM 3B 4e1t](https://huggingface.co/stabilityai/stablelm-3b-4e1t/). It is trained further on TinyCoT, but also on - 10000 comments from [reddit-instruct](https://hf.co/euclaise/reddit-instruct) - 15000 comments from [writingprompts-curated](https://hf.co/euclaise/writingprompts-curated) - 4000 examples of [converting MathQA problems to Python snippets](https://hf.co/euclaise/mathqa_programs) - 3000 examples of [shorter booksum cases (both chapter->summary and summary->chapter tasks)](https://huggingface.co/datasets/pszemraj/booksum-short) - 10000 examples from [mathoverflow-accepted](https://hf.co/euclaise/mathoverflow-accepted) comments with >10 upvotes - 1000 examples from [coedit-reworded-deduped](https://huggingface.co/datasets/BEE-spoke-data/coedit-reworded-deduped)