luna-4k-mini-ollama / README.md
marlonsousa's picture
Update README.md
1979bb6 verified
---
language:
- pt
- en
metrics:
- accuracy
pipeline_tag: question-answering
tags:
- personal
license: mit
---
# Luna Model
This document describes the use and functionalities of the personal model named Luna, which was trained based on the Phi-3 model. This model was developed for specific tasks as detailed below.
## Table of Contents
- [Introduction](#introduction)
- [Requirements](#requirements)
- [Installation](#installation)
- [Usage](#usage)
- [Features](#features)
- [Contribution](#contribution)
- [License](#license)
## Introduction
The Luna Model is a customized version of the Phi-3 model tailored for specific tasks such as text generation. This model leverages the capabilities of the Phi-3 architecture to provide efficient and accurate results for various natural language processing tasks.
## Requirements
- Ollama
# Installation
## Install Ollama
```shell
curl -fsSL https://ollama.com/install.sh | sh
```
# Usage
## Create Modelfile
```shell
touch Modelfile
```
## Modelfile content
```
FROM ./models/luna-4b-v0.5.gguf
PARAMETER temperature 1
"""
```
## Load the model
```bash
ollama create luna -f ./Modelfile
```
## Run Model
```bash
ollama run luna
```
# Usage Python
```python
import ollama
stream = ollama.chat(
model='llama3',
messages=[{'role': 'user', 'content': 'Who are you?'}],
stream=True,
)
for chunk in stream:
print(chunk['message']['content'], end='', flush=True)
```