10 epoch
Browse files- README.md +12 -44
- mergekit_config.yml +3 -3
- model-00001-of-00002.safetensors +1 -1
- model-00002-of-00002.safetensors +1 -1
README.md
CHANGED
@@ -37,8 +37,8 @@ Since [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-j
|
|
37 |
|
38 |
Using the [gemma-2-2b base model](https://huggingface.co/google/gemma-2-2b), I employed the ORPO method described by [mlabonne](https://towardsdatascience.com/fine-tune-llama-3-with-orpo-56cfab2f9ada) but the input model was read into VRAM by [unsloth](https://github.com/unslothai/unsloth) to allow using the full 40k dataset to run on a single 3090.
|
39 |
|
40 |
-
Five epoches was run. Smallest eval_loss was achieve at epoch
|
41 |
-
Checkpoint at epoch
|
42 |
applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18).
|
43 |
|
44 |
| Epoch | loss | eval_loss | eval_logps/rejected | eval_logps/chosen |
|
@@ -49,6 +49,12 @@ applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemm
|
|
49 |
| 4.00 | 1.5293 | 1.0166 | -1.2004 | -0.7200 |
|
50 |
| 4.96 | 1.2893 | 1.0077 | -1.1754 | -0.7106 |
|
51 |
| 5.00 | 1.3458 | 1.0078 | -1.1730 | -0.7105 |
|
|
|
|
|
|
|
|
|
|
|
|
|
52 |
|
53 |
Then I followed Rombodawg's [suggestion](https://www.reddit.com/r/LocalLLaMA/comments/1fyx27y/im_pretty_happy_with_how_my_method_worked_out/) to merge [gemma-2-2b](https://huggingface.co/google/gemma-2-2b), [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18) and [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain this model.
|
54 |
|
@@ -61,52 +67,14 @@ Click on the model name go to the raw score json generated by Open LLM Leaderboa
|
|
61 |
| Model | Average | IFEval | BHH | Math Lv5 | GPQA | MUSR | MMLU-PRO |
|
62 |
| ----- | ------- | ------ | ----|--------- | ---- | ---- | -------- |
|
63 |
| [gemma-2-2b-jpn-it](https://huggingface.co/datasets/open-llm-leaderboard/results/blob/main/google/gemma-2-2b-jpn-it/results_2024-10-15T15-21-39.173019.json) | 30.82 | 54.11 | 41.43 | 0.0 | 27.52 | 37.17 | 24.67 |
|
64 |
-
| gemma-2-2b-ORPO-jpn-it-abliterated-18 (5 epoches) |
|
65 |
-
| gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (
|
|
|
|
|
66 |
| [gemma-2-2b-jpn-it-abliterated-17](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-17/results_2024-10-18T15-18-46.821674.json) | 30.29 | 52.65 | 40.46 | 0.0 | 27.18 | 36.90 | 24.55 |
|
67 |
| [gemma-2-2b-jpn-it-abliterated-18](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-18/results_2024-10-18T15-41-42.399571.json) | 30.61 | 53.02 | 40.96 | 0.0 | 27.35 | 37.30 | 25.05 |
|
68 |
| [gemma-2-2b-jpn-it-abliterated-24](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-24/results_2024-10-25T16-29-46.542899.json) | 30.61 | 51.37 | 40.77 | 0.0 | 27.77 | 39.02 | 24.73 |
|
69 |
|
70 |
-
## Merge Details
|
71 |
-
### Merge Method
|
72 |
-
|
73 |
-
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
74 |
-
|
75 |
-
This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [google/gemma-2-2b](https://huggingface.co/google/gemma-2-2b) as a base.
|
76 |
-
|
77 |
-
### Models Merged
|
78 |
-
|
79 |
-
The following models were included in the merge:
|
80 |
-
* ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
|
81 |
-
* ymcki/gemma-2-2b-jpn-it-abliterated-18
|
82 |
-
|
83 |
-
### Configuration
|
84 |
-
|
85 |
-
The following YAML configuration was used to produce this model:
|
86 |
-
|
87 |
-
```yaml
|
88 |
-
models:
|
89 |
-
- model: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
|
90 |
-
dtype: bfloat16
|
91 |
-
parameters:
|
92 |
-
density: 1.0
|
93 |
-
weight: 1.0
|
94 |
-
- model: ymcki/gemma-2-2b-jpn-it-abliterated-18
|
95 |
-
dtype: bfloat16
|
96 |
-
parameters:
|
97 |
-
density: 1.0
|
98 |
-
weight: 1.0
|
99 |
-
merge_method: ties
|
100 |
-
base_model: google/gemma-2-2b
|
101 |
-
parameters:
|
102 |
-
density: 1.0
|
103 |
-
weight: 1.0
|
104 |
-
normalize: true
|
105 |
-
int8_mask: true
|
106 |
-
dtype: bfloat16
|
107 |
-
tokenizer_source: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
|
108 |
-
```
|
109 |
-
|
110 |
## How to run this model
|
111 |
|
112 |
```py
|
|
|
37 |
|
38 |
Using the [gemma-2-2b base model](https://huggingface.co/google/gemma-2-2b), I employed the ORPO method described by [mlabonne](https://towardsdatascience.com/fine-tune-llama-3-with-orpo-56cfab2f9ada) but the input model was read into VRAM by [unsloth](https://github.com/unslothai/unsloth) to allow using the full 40k dataset to run on a single 3090.
|
39 |
|
40 |
+
Five epoches was run. Smallest eval_loss was achieve at epoch 7.72.
|
41 |
+
Checkpoint at epoch 7.72 is used to obtain a model adapter and
|
42 |
applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18).
|
43 |
|
44 |
| Epoch | loss | eval_loss | eval_logps/rejected | eval_logps/chosen |
|
|
|
49 |
| 4.00 | 1.5293 | 1.0166 | -1.2004 | -0.7200 |
|
50 |
| 4.96 | 1.2893 | 1.0077 | -1.1754 | -0.7106 |
|
51 |
| 5.00 | 1.3458 | 1.0078 | -1.1730 | -0.7105 |
|
52 |
+
| 6.00 | 1.3807 | 0.9924 | -1.1757 | -0.6971 |
|
53 |
+
| 7.00 | 1.0855 | 0.9889 | -1.2634 | -0.7235 |
|
54 |
+
| 7.72 | 0.8720 | 0.9855 | -1.2374 | -0.7100 |
|
55 |
+
| 8.00 | 0.7301 | 0.9864 | -1.2406 | -0.7113 |
|
56 |
+
| 9.00 | 1.1939 | 0.9934 | -1.2703 | -0.6852 |
|
57 |
+
| 10.00 | 0.7421 | 1.0269 | -1.2552 | -0.7395 |
|
58 |
|
59 |
Then I followed Rombodawg's [suggestion](https://www.reddit.com/r/LocalLLaMA/comments/1fyx27y/im_pretty_happy_with_how_my_method_worked_out/) to merge [gemma-2-2b](https://huggingface.co/google/gemma-2-2b), [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18) and [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain this model.
|
60 |
|
|
|
67 |
| Model | Average | IFEval | BHH | Math Lv5 | GPQA | MUSR | MMLU-PRO |
|
68 |
| ----- | ------- | ------ | ----|--------- | ---- | ---- | -------- |
|
69 |
| [gemma-2-2b-jpn-it](https://huggingface.co/datasets/open-llm-leaderboard/results/blob/main/google/gemma-2-2b-jpn-it/results_2024-10-15T15-21-39.173019.json) | 30.82 | 54.11 | 41.43 | 0.0 | 27.52 | 37.17 | 24.67 |
|
70 |
+
| [gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (5 epoches)](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge/results_2024-10-30T17-06-58.119904.json) | 29.26 | 49.16 | 38.15 | 2.49 | 28.19 | 33.07 | 24.51 |
|
71 |
+
| gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (10 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
|
72 |
+
| [gemma-2-2b-ORPO-jpn-it-abliterated-18 (5 epoches)](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18/results_2024-10-30T22-19-29.202883.json) | 29.57 | 48.05 | 41.26 | 0.0 | 27.18 | 36.51 | 24.43
|
73 |
+
| gemma-2-2b-ORPO-jpn-it-abliterated-18 (10 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
|
74 |
| [gemma-2-2b-jpn-it-abliterated-17](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-17/results_2024-10-18T15-18-46.821674.json) | 30.29 | 52.65 | 40.46 | 0.0 | 27.18 | 36.90 | 24.55 |
|
75 |
| [gemma-2-2b-jpn-it-abliterated-18](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-18/results_2024-10-18T15-41-42.399571.json) | 30.61 | 53.02 | 40.96 | 0.0 | 27.35 | 37.30 | 25.05 |
|
76 |
| [gemma-2-2b-jpn-it-abliterated-24](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-24/results_2024-10-25T16-29-46.542899.json) | 30.61 | 51.37 | 40.77 | 0.0 | 27.77 | 39.02 | 24.73 |
|
77 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
78 |
## How to run this model
|
79 |
|
80 |
```py
|
mergekit_config.yml
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
models:
|
2 |
-
- model:
|
3 |
dtype: bfloat16
|
4 |
parameters:
|
5 |
density: 1.0
|
6 |
weight: 1.0
|
7 |
-
- model:
|
8 |
dtype: bfloat16
|
9 |
parameters:
|
10 |
density: 1.0
|
@@ -17,4 +17,4 @@ parameters:
|
|
17 |
normalize: true
|
18 |
int8_mask: true
|
19 |
dtype: bfloat16
|
20 |
-
tokenizer_source:
|
|
|
1 |
models:
|
2 |
+
- model: ./gemma-2-2b-ORPO-jpn-it-abliterated-18
|
3 |
dtype: bfloat16
|
4 |
parameters:
|
5 |
density: 1.0
|
6 |
weight: 1.0
|
7 |
+
- model: /home/user/gemma-2-2b-jpn-it-abliterated-18
|
8 |
dtype: bfloat16
|
9 |
parameters:
|
10 |
density: 1.0
|
|
|
17 |
normalize: true
|
18 |
int8_mask: true
|
19 |
dtype: bfloat16
|
20 |
+
tokenizer_source: ./gemma-2-2b-ORPO-jpn-it-abliterated-18
|
model-00001-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4959727696
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:44f62f5fe1d112022beb697189fa7de24d883141a78d8e4cb2c013602c39ac45
|
3 |
size 4959727696
|
model-00002-of-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 268999016
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c66000edb5b2bd0d28c2aafea08c18cd7324d956e6daca3b5196dc199027492
|
3 |
size 268999016
|