Model Card for Model ID

Merged and GPTQ quantized version of rayliuca/TRagx-internlm2-7b

Note: I'm having some difficulties quantizing the models using GPTQ. Mistral and NeuralOmniBeagle's GPTQ models have significantly degraded output, while quantized TowerInstruct v0.2 was not working out right

While this quantized model for InternLM2 seems to work all right, the translation accuracy is not validated.

These AWQ quantized models are recommended:

GPTQ Dataset

Qutanized with nsamples=45 * 3 languages [ja, zh, en] from the c4 dataset

License

See the original InternLM2 repo https://huggingface.co/internlm/internlm2-7b#open-source-license

Downloads last month
5
Safetensors
Model size
1.69B params
Tensor type
I32
·
FP16
·
Inference Examples
Inference API (serverless) does not yet support model repos that contain custom code.

Collection including rayliuca/TRagx-GPTQ-internlm2-7b