Transformers
GGUF
llama
File size: 1,889 Bytes
f89b70e
 
 
08b4a41
 
 
 
 
67d51f0
08b4a41
67d51f0
08b4a41
1ce4968
08b4a41
 
 
1ce4968
67d51f0
 
08b4a41
1ce4968
67d51f0
1ce4968
 
 
08b4a41
 
 
 
67d51f0
 
08b4a41
 
 
 
 
 
 
 
 
 
67d51f0
08b4a41
 
 
 
 
1ce4968
08b4a41
1ce4968
08b4a41
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
license: apache-2.0
---

# Model Card for Model ID

<!-- Provide a quick summary of what the model is/does. -->

**slim-sql-tool** is part of the SLIM ("Structured Language Instruction Model") model series, providing a set of small, specialized decoder-based LLMs, fine-tuned for function-calling.  

slim-sql-tool is a 4_K_M quantized GGUF version of slim-sql, providing a small, fast inference implementation.

Load in your favorite GGUF inference engine (see details in config.json to set up the prompt template), or try with llmware as follows:

    from llmware.models import ModelCatalog

    # to load the model and make a basic inference
    sql_tool = ModelCatalog().load_model("slim-sql-tool")
    response = sql_tool.function_call(text_sample)  

    # this one line will download the model and run a series of tests
    ModelCatalog().test_run("slim-sql-tool", verbose=True)  
    
    
Slim models can also be loaded even more simply as part of a multi-model, multi-step LLMfx calls:

    from llmware.agents import LLMfx

    llm_fx = LLMfx()
    llm_fx.load_tool("sql")
    response = llm_fx.sql(text)


### Model Description

<!-- Provide a longer summary of what this model is. -->

- **Developed by:** llmware
- **Model type:** GGUF 
- **Language(s) (NLP):** English
- **License:** Apache 2.0
- **Quantized from model:** llmware/slim-sql (finetuned tiny llama)

## Uses

<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->

SLIM models provide a fast, flexible, intuitive way to integrate classifiers and structured function calls into RAG and LLM application workflows.  

Model instructions, details and test samples have been packaged into the config.json file in the repository, along with the GGUF file.   



## Model Card Contact

Darren Oberst & llmware team