This is an ExLlamaV2 quantized model in 3bpw of jebcarter/Psyfighter-13B using PIPPA as the calibration dataset. (I might reupload this or add a new branch with a different version that uses the default calibration dataset instead)

Alpaca and ChatML prompt templates seem to work fine with this model.

Original Model card

merge_method: task_arithmaetic
base_model: TheBloke/Llama-2-13B-fp16
models:
  - model: TheBloke/Llama-2-13B-fp16
  - model: KoboldAI/LLaMA2-13B-Tiefighter
    parameters:
      weight: 1.0
  - model: chaoyi-wu/MedLLaMA_13B
    parameters:
      weight: 0.01
  - model: Doctor-Shotgun/llama-2-13b-chat-limarp-v2-merged
    parameters:
      weight: 0.02
dtype: float16

This model was made possible thanks to the Compute provided by the KoboldAI community.

Downloads last month
17
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Collection including mpasila/Psyfighter-13B-exl2-3bpw