File size: 2,042 Bytes
db6a0be
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3a7d7a5
 
3e3bd97
f2a12fa
 
3e3bd97
230978a
3e3bd97
6b0b6c6
c1c398d
7f0471c
 
981c056
 
db6a0be
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
base_model:
- inflatebot/thorn-0.5
- inflatebot/thorn-0.35
- inflatebot/thorn-0.55
- inflatebot/thorn-0.45
library_name: transformers
tags:
- mergekit
- merge

---
# merge

This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).

## Merge Details
![Made with NovelAI](https://huggingface.co/inflatebot/guns-and-roses-r1/resolve/main/guns%20and%20roses.png)
`Quickest draw in the West.`

### NOTE: If you are getting phrase repetition or nonsense outputs with SillyTavern, make sure that "Include names" is disabled under Advanced Formatting. Nemo models tend to exhibit these issues when this is enabled.

A re-application of the Helium-3 process to Mistral Nemo analogues. Experimental (as you can tell by it having a revision number, I'll be playing with this more in coming time.)
Based ultimately on [Magnum-12B-V2](https://huggingface.co/anthracite-org/magnum-12b-v2) and [MN-12B-Rosier-v1](https://huggingface.co/Fizzarolli/MN-12b-Rosier-v1).

Quants available from [Reiterate3680](https://huggingface.co/Reiterate3680/guns-and-roses-r1-GGUF/tree/main)

Special thanks to Fizz and Toasty Pigeon.

Use **ChatML** formatting, as Rosier was trained from base (so no instruct format) and Magnum V2 was trained on ChatML!

### Merge Method

This model was merged using the [Model Stock](https://arxiv.org/abs/2403.19522) merge method using [inflatebot/thorn-0.35](https://huggingface.co/inflatebot/thorn-0.35) as a base.

### Models Merged

The following models were included in the merge:
* [inflatebot/thorn-0.5](https://huggingface.co/inflatebot/thorn-0.5)
* [inflatebot/thorn-0.55](https://huggingface.co/inflatebot/thorn-0.55)
* [inflatebot/thorn-0.45](https://huggingface.co/inflatebot/thorn-0.45)

### Configuration

The following YAML configuration was used to produce this model:

```yaml
models:
  - model: inflatebot/thorn-0.5
  - model: inflatebot/thorn-0.45
  - model: inflatebot/thorn-0.55

merge_method: model_stock
base_model: inflatebot/thorn-0.35
dtype: bfloat16
```