Introduction

The model is trained with Masked thought Fine-Tuning (MFT), a simple variant of standard Supervised Fine-Tuning (SFT). You can refer to our code and paper below.

Links

Results

We test it with the scripts provided in our code.

Downloads last month
12
Safetensors
Model size
6.74B params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Dataset used to train adalaw/Llama2-7B-GSM8K-MFT