|
--- |
|
library_name: transformers |
|
tags: |
|
- quantization |
|
- phi-3 |
|
license: apache-2.0 |
|
language: |
|
- en |
|
--- |
|
|
|
# Model Card for Model ID |
|
A 4-bits double-quantized version of [ernestoBocini/Phi3-mini-DPO-Tuned](https://huggingface.co/ernestoBocini/Phi3-mini-DPO-Tuned). |
|
|
|
## Model Details |
|
|
|
This is a Phi-3-mini-4k-instruct fine-tuned with SFT and DPO on STEM domains, and finally quantized to a 4 bits precision, to serve the purpose of being an AI university tutor. |
|
|
|
Quantization config used: |
|
```python |
|
BitsAndBytesConfig( |
|
load_in_4bit=True, |
|
bnb_4bit_quant_type="nf4", |
|
bnb_4bit_use_double_quant=True, |
|
bnb_4bit_compute_dtype=torch.bfloat16 |
|
) |
|
``` |