|
# Data query example |
|
|
|
Example of integration with HuggingFace Inference API with help of [langchaingo](https://github.com/tmc/langchaingo). |
|
|
|
## Setup |
|
|
|
Download the LocalAI and start the API: |
|
|
|
```bash |
|
# Clone LocalAI |
|
git clone https://github.com/go-skynet/LocalAI |
|
|
|
cd LocalAI/examples/langchain-huggingface |
|
|
|
docker-compose up -d |
|
``` |
|
|
|
Node: Ensure you've set `HUGGINGFACEHUB_API_TOKEN` environment variable, you can generate it |
|
on [Settings / Access Tokens](https://huggingface.co/settings/tokens) page of HuggingFace site. |
|
|
|
This is an example `.env` file for LocalAI: |
|
|
|
```ini |
|
MODELS_PATH=/models |
|
CONTEXT_SIZE=512 |
|
HUGGINGFACEHUB_API_TOKEN=hg_123456 |
|
``` |
|
|
|
## Using remote models |
|
|
|
Now you can use any remote models available via HuggingFace API, for example let's enable using of |
|
[gpt2](https://huggingface.co/gpt2) model in `gpt-3.5-turbo.yaml` config: |
|
|
|
```yml |
|
name: gpt-3.5-turbo |
|
parameters: |
|
model: gpt2 |
|
top_k: 80 |
|
temperature: 0.2 |
|
top_p: 0.7 |
|
context_size: 1024 |
|
backend: "langchain-huggingface" |
|
stopwords: |
|
- "HUMAN:" |
|
- "GPT:" |
|
roles: |
|
user: " " |
|
system: " " |
|
template: |
|
completion: completion |
|
chat: gpt4all |
|
``` |
|
|
|
Here is you can see in field `parameters.model` equal `gpt2` and `backend` equal `langchain-huggingface`. |
|
|
|
## How to use |
|
|
|
```shell |
|
# Now API is accessible at localhost:8080 |
|
curl http://localhost:8080/v1/models |
|
# {"object":"list","data":[{"id":"gpt-3.5-turbo","object":"model"}]} |
|
|
|
curl http://localhost:8080/v1/completions -H "Content-Type: application/json" -d '{ |
|
"model": "gpt-3.5-turbo", |
|
"prompt": "A long time ago in a galaxy far, far away", |
|
"temperature": 0.7 |
|
}' |
|
``` |