metadata
base_model:
- euclaise/Memphis-CoT-3B
license: cc-by-4.0
datasets:
- euclaise/TinyCoT
- euclaise/mathoverflow-accepted
- euclaise/reddit-instruct
- euclaise/WritingPrompts_curated
- sablo/oasst2_curated
- euclaise/mathqa_programs
- BEE-spoke-data/coedit-reworded-deduped
- pszemraj/booksum-short
library_name: transformers
tags:
- supertrainer2000
Memphis-scribe 3B is a finetune of Memphis-CoT 3B on more creative data, which itself is a finetune of StableLM 3B 4e1t.
It is trained further on TinyCoT, but also on
- 10000 comments from reddit-instruct
- 15000 comments from writingprompts-curated
- 4000 examples of converting MathQA problems to Python snippets
- 3000 examples of shorter booksum cases (both chapter->summary and summary->chapter tasks)
- 10000 examples from mathoverflow-accepted comments with >10 upvotes
- 1000 examples from coedit-reworded-deduped