whoami02's picture
Update README.md
01603e1 verified
---
license: apache-2.0
language:
- en
tags:
- HuggingFace
- defog/sqlcoder.gguf
- sqlcoder-7b-2.gguf
pipeline_tag: text-generation
---
## Model Details
I do not claim ownership of this model. <br>
It is converted into 8-bit GGUF format from original repository [huggingface.co/defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2)
### Model Description
**Developed by:** [Defog AI](https://defog.ai)
### Model Sources
**Repository:** [https://huggingface.co/defog/sqlcoder-7b-2](https://huggingface.co/defog/sqlcoder-7b-2)
### Example usage
**With Llamacpp:**
```
from langchain_community.llms. llamacpp import Llamacpp
from huggingface_hub import hf_hub_download
YOUR_MODEL_DIRECTORY = None
CONTEXT LENGHT = None
MAX TOKENS = None
BATCH SIZE = None
TEMPERATURE = None
GPU_OFFLOAD = None
def load_model (model_id, model_basename):
model_path = hf_hub_download (
repo_id=model_id,
filename=model_basename,
resume_download=True,
cache_dir="YOUR_MODEL_DIRECTORY",
)
kwargs = {
'model_path': model_path,
'n_ctx': CONTEXT_LENGHT,
'max_tokens': MAX_TOKENS,
'n_batch': BATCH_SIZE,
'n_gpu_layers': GPU_OFFLOAD,
'temperature': TEMPERATURE,
'verbose': True,
}
return LlamaCpp(**kwargs)
11m = load_model(
model_id="whoami02/defog-sqlcoder-2-GGUF",
model_basename="sqlcoder-7b-2.q8_0.gguf",
```
<!-- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/64e4b2a292418458795a5cb5/hX-xeG1VKwIbumDGNT05W.png) -->