|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
tags: |
|
- HuggingFace |
|
- defog/sqlcoder.gguf |
|
- sqlcoder-7b-2.gguf |
|
pipeline_tag: text-generation |
|
--- |
|
## Model Details |
|
I do not claim ownership of this model. <br> |
|
It is converted into 8-bit GGUF format from original repository [huggingface.co/defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2) |
|
|
|
### Model Description |
|
|
|
**Developed by:** [Defog AI](https://defog.ai) |
|
|
|
### Model Sources |
|
|
|
**Repository:** [https://huggingface.co/defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2) |
|
|
|
### Example usage |
|
**With Llamacpp:** |
|
``` |
|
from langchain_community.llms. llamacpp import Llamacpp |
|
from huggingface_hub import hf_hub_download |
|
|
|
YOUR_MODEL_DIRECTORY = None |
|
CONTEXT LENGHT = None |
|
MAX TOKENS = None |
|
BATCH SIZE = None |
|
TEMPERATURE = None |
|
GPU_OFFLOAD = None |
|
|
|
def load_model (model_id, model_basename): |
|
model_path = hf_hub_download ( |
|
repo_id=model_id, |
|
filename=model_basename, |
|
resume_download=True, |
|
cache_dir="YOUR_MODEL_DIRECTORY", |
|
) |
|
kwargs = { |
|
'model_path': model_path, |
|
'n_ctx': CONTEXT_LENGHT, |
|
'max_tokens': MAX_TOKENS, |
|
'n_batch': BATCH_SIZE, |
|
'n_gpu_layers': GPU_OFFLOAD, |
|
'temperature': TEMPERATURE, |
|
'verbose': True, |
|
} |
|
return LlamaCpp(**kwargs) |
|
|
|
11m = load_model( |
|
model_id="whoami02/defog-sqlcoder-2-GGUF", |
|
model_basename="sqlcoder-7b-2.q8_0.gguf", |
|
``` |
|
<!--  --> |