|
--- |
|
base_model: unsloth/DeepSeek-R1-Distill-Qwen-7B-unsloth-bnb-4bit |
|
library_name: transformers |
|
license: apache-2.0 |
|
datasets: |
|
- leonvanbokhorst/friction-uncertainty-v2 |
|
language: |
|
- en |
|
tags: |
|
- ai-safety |
|
- ai-friction |
|
- human-like-messiness |
|
- ai-uncertainty |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# Friction Reasoning Model |
|
|
|
This model is fine-tuned to respond in an uncertain manner. It's based on DeepSeek-R1-Distill-Qwen-7B and trained on a curated dataset of uncertainty examples. |
|
|
|
## Model Description |
|
|
|
- **Model Architecture**: DeepSeek-R1-Distill-Qwen-7B with LoRA adapters |
|
- **Language(s)**: English |
|
- **License**: Apache 2.0 |
|
- **Finetuning Approach**: Instruction tuning with friction-based reasoning examples |
|
|
|
|
|
### Limitations |
|
|
|
The model: |
|
- Is not designed for factual question-answering |
|
- May sometimes be overly uncertain |
|
- Should not be used for medical, legal, or financial advice |
|
- May not perform well on objective or factual tasks |
|
|
|
### Bias and Risks |
|
|
|
The model: |
|
- May exhibit biases present in the training data |
|
- Could potentially reinforce uncertainty in certain situations |
|
- Might challenge user assumptions in sensitive contexts |
|
- Should be used with appropriate content warnings |
|
|
|
|
|
|
|
|
|
## Citation |
|
|
|
If you use this model in your research, please cite: |
|
|
|
```bibtex |
|
@misc{friction-reasoning-2025, |
|
author = {Leon van Bokhorst}, |
|
title = {Mixture of Friction: Fine-tuned Language Model for Uncertainty}, |
|
year = {2025}, |
|
publisher = {HuggingFace}, |
|
journal = {HuggingFace Model Hub}, |
|
howpublished = {\url{https://huggingface.co/leonvanbokhorst/deepseek-r1-uncertainty}} |
|
} |
|
``` |
|
|
|
## Acknowledgments |
|
|
|
- DeepSeek AI for the base model |
|
- Unsloth team for the optimization toolkit |
|
- HuggingFace for the model hosting and infrastructure |