|
---
|
|
language:
|
|
- pt
|
|
- en
|
|
metrics:
|
|
- accuracy
|
|
pipeline_tag: question-answering
|
|
tags:
|
|
- personal
|
|
license: mit
|
|
---
|
|
|
|
# Luna Model |
|
|
|
This document describes the use and functionalities of the personal model named Luna, which was trained based on the Phi-3 model. This model was developed for specific tasks as detailed below. |
|
|
|
## Table of Contents |
|
- [Introduction](#introduction) |
|
- [Requirements](#requirements) |
|
- [Installation](#installation) |
|
- [Usage](#usage) |
|
- [Features](#features) |
|
- [Contribution](#contribution) |
|
- [License](#license) |
|
|
|
## Introduction |
|
The Luna Model is a customized version of the Phi-3 model tailored for specific tasks such as text generation. This model leverages the capabilities of the Phi-3 architecture to provide efficient and accurate results for various natural language processing tasks. |
|
|
|
## Requirements |
|
- Ollama |
|
|
|
# Installation |
|
|
|
## Install Ollama |
|
|
|
```shell |
|
curl -fsSL https://ollama.com/install.sh | sh |
|
``` |
|
|
|
# Usage |
|
|
|
## Create Modelfile |
|
|
|
```shell |
|
touch Modelfile |
|
``` |
|
|
|
## Modelfile content |
|
|
|
``` |
|
FROM ./models/luna-4b-v0.5.gguf |
|
|
|
PARAMETER temperature 1 |
|
""" |
|
``` |
|
|
|
## Load the model |
|
|
|
```bash |
|
ollama create luna -f ./Modelfile |
|
``` |
|
|
|
## Run Model |
|
|
|
```bash |
|
ollama run luna |
|
``` |
|
|
|
# Usage Python |
|
|
|
```python |
|
import ollama |
|
|
|
stream = ollama.chat( |
|
model='llama3', |
|
messages=[{'role': 'user', 'content': 'Who are you?'}], |
|
stream=True, |
|
) |
|
|
|
for chunk in stream: |
|
print(chunk['message']['content'], end='', flush=True) |
|
``` |