appvoid/palmer-x-002-GGUF

Quantized GGUF model files for palmer-x-002 from appvoid

Name Quant method Size
palmer-x-002.fp16.gguf fp16 2.20 GB
palmer-x-002.q2_k.gguf q2_k 483.12 MB
palmer-x-002.q3_k_m.gguf q3_k_m 550.82 MB
palmer-x-002.q4_k_m.gguf q4_k_m 668.79 MB
palmer-x-002.q5_k_m.gguf q5_k_m 783.02 MB
palmer-x-002.q6_k.gguf q6_k 904.39 MB
palmer-x-002.q8_0.gguf q8_0 1.17 GB

Original Model Card:

palmer

x-002

This is an incremental model update on palmer-002 using dpo technique. X means dpo+sft spinoff.

evaluation

Model ARC_C HellaSwag PIQA Winogrande
tinyllama-2t 0.2807 0.5463 0.7067 0.5683
palmer-001 0.2807 0.5524 0.7106 0.5896
tinyllama-2.5t 0.3191 0.5896 0.7307 0.5872
palmer-002 0.3242 0.5956 0.7345 0.5888
palmer-x-002 0.3224 0.5941 0.7383 0.5912

training

~500 dpo samples as experimental data to check on improvements. It seems like data is making it better on some benchmarks while also degrading quality on others.

prompt

no prompt

As you can notice, the model actually completes by default questions that are the most-likely to be asked, which is good because most people will use it to answer as a chatbot. Buy Me A Coffee

Downloads last month
24
GGUF
Model size
1.1B params
Architecture
llama

2-bit

4-bit

5-bit

6-bit

8-bit

Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for afrideva/palmer-x-002-GGUF

Quantized
(1)
this model

Dataset used to train afrideva/palmer-x-002-GGUF