The Rho-1 series are pretrained language models that utilize Selective Language Modeling (SLM) objectives.
In math reasoning pretraining, SLM improves average few-shot accuracy on GSM8k and MATH by over 16%, achieving the baseline performance 5-10x faster.
For more details please check our github and paper.