Overview

This is a bare model without any output layer or classification head. It has been quantized to be used for feature extraction tasks.

Usage

This model is intended to be used as a base for training on downstream tasks. In order to use it for predictions and inference, it should be fine-tuned on a specific task with an appropriate output layer or classification head added.

Quantization

The model has been quantized using the following parameters:

Lora alpha: 16

Lora rank: 32

Lora target modules: all-linear

bits: 4

LoftQ iterations: 5

Downloads last month
7
Safetensors
Model size
7.5B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.