metadata
library_name: peft
Training procedure
The following bitsandbytes
quantization config was used during training:
- bits: 4
- checkpoint_format: gptq
- desc_act: True
- dynamic: None
- group_size: 128
- lm_head: False
- meta: {'damp_auto_increment': 0.0015, 'damp_percent': 0.01, 'quantizer': ['gptqmodel:1.4.0-dev'], 'static_groups': False, 'true_sequential': True, 'uri': 'https://github.com/modelcloud/gptqmodel'}
- quant_method: gptq
- sym: True
Framework versions
- PEFT 0.5.0