TinyLlama_v1.1_1bit_BitDistiller
This is a 1-bit quantized version of TinyLlama v1.1, trained using BitDistiller with asymmetric quantization and self-distillation (CAKLD) to optimize accuracy retention under extreme compression. The model is fine-tuned on WikiText-2 and Alpaca-cleaned datasets and evaluated on multiple-choice QA benchmarks.
Key Features:
- 1-bit quantization for ultra-efficient inference.
- Asymmetric weight clipping to reduce precision loss.
- CAKLD knowledge distillation to preserve performance.
- Tested on ARC-Challenge, HellaSwag, PIQA, and Winogrande.
- Downloads last month
- 27
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
HF Inference deployability: The model has no library tag.
Model tree for fredericowieser/TinyLlama_v1.1_mix_wikitext_alpaca_1bit_BitDistiller_baseline
Base model
TinyLlama/TinyLlama_v1.1Datasets used to train fredericowieser/TinyLlama_v1.1_mix_wikitext_alpaca_1bit_BitDistiller_baseline
Evaluation results
- Accuracy on ARC-Challengetest set self-reported0.215
- Normalized Accuracy on ARC-Challengetest set self-reported0.247
- Accuracy on HellaSwagtest set self-reported0.257
- Normalized Accuracy on HellaSwagtest set self-reported0.253
- Accuracy on PIQAvalidation set self-reported0.528
- Normalized Accuracy on PIQAvalidation set self-reported0.503
- Accuracy on Winograndetest set self-reported0.512
- QA Average on QA-Avgself-reported0.378