|
--- |
|
library_name: transformers |
|
tags: |
|
- transformers |
|
- peft |
|
- arxiv:2406.08391 |
|
license: llama2 |
|
base_model: meta-llama/Llama-2-7b-hf |
|
datasets: |
|
- calibration-tuning/Llama-2-7b-hf-20k-oe |
|
--- |
|
|
|
# Model Card |
|
|
|
**Llama 2 7B CT-OE** is a fine-tuned [Llama 2 7B](https://huggingface.co/meta-llama/Llama-2-7b-hf) model that provides well-calibrated confidence estimates for open-ended question answering. |
|
|
|
The model is fine-tuned (calibration-tuned) using a [dataset](https://huggingface.co/datasets/calibration-tuning/Llama-2-7b-hf-20k-oe) of *open-ended* generations from `meta-llama/Llama-2-7b-hf`, labeled for correctness. |
|
At test/inference time, the probability of correctness defines the confidence of the model in its answer. |
|
For full details, please see our [paper](https://arxiv.org/abs/2406.08391) and supporting [code](https://github.com/activatedgeek/calibration-tuning). |
|
|
|
**Other Models**: We also release a broader collection of [Open-Ended CT Models](https://huggingface.co/collections/calibration-tuning/open-ended-ct-models-66043b12c7902115c826a20e). |
|
|
|
## Usage |
|
|
|
This adapter model is meant to be used on top of `meta-llama/Llama-2-7b-hf` model generations. |
|
|
|
The confidence estimation pipeline follows these steps, |
|
1. Load base model and PEFT adapter. |
|
2. Disable adapter and generate answer. |
|
3. Enable adapter and generate confidence. |
|
|
|
All standard guidelines for the base model's generation apply. |
|
|
|
For a complete example, see [play.py](https://github.com/activatedgeek/calibration-tuning/blob/main/experiments/play.py) at the supporting code repository. |
|
|
|
**NOTE**: Using the adapter for generations may hurt downstream task accuracy and confidence estimates. We recommend using the adapter to estimate *only* confidence. |
|
|
|
## License |
|
|
|
The model is released under the original model's Llama 2 Community License Agreement. |