File size: 3,689 Bytes
04f5373
 
 
 
 
008f864
04f5373
 
0157d20
ffff1f8
 
 
 
 
4743f67
ffff1f8
 
 
 
 
 
c87cebd
 
 
 
ffff1f8
c87cebd
 
ffff1f8
 
 
 
4743f67
ffff1f8
052e729
 
ffff1f8
 
 
 
 
4743f67
ffff1f8
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4743f67
ffff1f8
ead8392
ffff1f8
b4fb33d
ffff1f8
93f262a
ffff1f8
b4fb33d
ffff1f8
93f262a
 
 
 
ffff1f8
 
 
 
 
 
 
 
 
 
 
4743f67
ffff1f8
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
---
license: apache-2.0
language:
- en
base_model:
- meta-llama/Llama-3.1-8B
pipeline_tag: text-generation
---
# Cat1.0

![Cover Image](https://i.ibb.co/PYCdt9n/3i-RPOp-Vn-Tb-O4-E021n6-Pljg.jpg)

## Overview

Cat1.0 is a fine-tuned version of **Llama-3-1-8b base model**, optimized for roleplaying, logic, and reasoning tasks. Utilizing iterative fine-tuning and human-AI chat logs, this model works well for numerous chat scenarios.

## Model Specifications

- **Parameters**: 8 Billion (8B)
- **Precision**: bf16 (Brain Floating Point 16-bit)
- **Fine-Tuning Method**: LoRa (Low-Rank Adaptation)
- **Lora Rank**: 32
- **Lora Alpha**: 64
- **Learning Rate**: 0.0008
- **Training Epochs**: 4
- **Datasets Used**:
  - cat1.0 Roleplay Dataset
  - cat1.0 Reasoning and Logic Dataset
- **Fine-Tuning Approach**: Iterative Fine-Tuning using self-chat logs

## Recommended Settings

To achieve optimal performance with this model, I recommend the following settings:

- **Temperature**: `1.1`
- **Min P**: `0.05`

> **Note**: Due to the nature of the fine-tuning, setting the temperature to `1.1` or higher helps prevent the model from repeating itself and encourages more creative and coherent responses.

## Usage Instructions

I recommend using the [oobabooga text-generation-webui](https://github.com/oobabooga/text-generation-webui) for an optimal experience. Load the model in `bf16` precision and enable `flash-attention2` for improved performance.

### Installation Steps

1. **Clone the WebUI Repository**:

   ```bash
   git clone https://github.com/oobabooga/text-generation-webui
   cd text-generation-webui
   ```

2. **Install Dependencies**:

   ```bash
   pip install -r requirements.txt
   ```

3. **Download the Model**:

   Download the fine-tuned model from [Hugging Face](#) and place it in the `models` directory.

4. **Launch the WebUI**:

   ```bash
   python server.py --bf16 --flash-attention
   ```

### Sample Prompt Formats

You can interact with the model using either **chat format** or **chat-instruct format**. Here's an example:

```plaintext
Ryan is a computer engineer who works at Intel.

Ryan: Hey, how's it going Natalie?
Natalie: Good, how are things going with you, Ryan?
Ryan: Great, I'm just doing just great.
```

## Model Capabilities

Below are some examples showcasing the model's performance in various roleplay scenarios:

### Roleplay Examples

![Roleplay Log 1](https://i.ibb.co/Zz20Wxw/Screenshot-46.png)

![Roleplay Log 2](https://i.ibb.co/wWrdsZm/Screenshot-49-1.png)

![Roleplay Log 3](https://i.ibb.co/4PG7W2K/Screenshot-47.png)

### Text Generation Example

![Text Generation Example](https://i.ibb.co/J5ZVCnR/Screenshot-45.png)

## Limitations and Tips

While this model excels in chat and roleplaying scenarios, it isn't perfect. If you notice the model repeating itself or providing less coherent responses:

- **Increase the Temperature**: Setting the temperature higher (≥ `1.1`) can help generate more diverse and creative outputs.
- **Adjust `min_p` Setting**: Ensuring `min_p` is at least `0.05` can prevent low-probability tokens from being excluded, enhancing the response quality.

## Acknowledgments

- **oobabooga text-generation-webui**: A powerful interface for running and interacting with language models. [GitHub Repository](https://github.com/oobabooga/text-generation-webui)
- **Hugging Face**: For hosting the model and providing a platform for collaboration. [Website](https://huggingface.co/)
- **Meta** For pre-training the Llama-3.1-8B Base Model that was used for fine-tuning. [Model Card](https://huggingface.co/meta-llama/Llama-3.1-8B)

*For any issues or questions, please open an issue in this repository.*