ymcki commited on
Commit
d1f1b61
·
1 Parent(s): a7e5414
README.md CHANGED
@@ -37,8 +37,8 @@ Since [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-j
37
 
38
  Using the [gemma-2-2b base model](https://huggingface.co/google/gemma-2-2b), I employed the ORPO method described by [mlabonne](https://towardsdatascience.com/fine-tune-llama-3-with-orpo-56cfab2f9ada) but the input model was read into VRAM by [unsloth](https://github.com/unslothai/unsloth) to allow using the full 40k dataset to run on a single 3090.
39
 
40
- Five epoches was run. Smallest eval_loss was achieve at epoch 4.96.
41
- Checkpoint at epoch 4.96 is used to obtain a model adapter and
42
  applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18).
43
 
44
  | Epoch | loss | eval_loss | eval_logps/rejected | eval_logps/chosen |
@@ -49,6 +49,12 @@ applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemm
49
  | 4.00 | 1.5293 | 1.0166 | -1.2004 | -0.7200 |
50
  | 4.96 | 1.2893 | 1.0077 | -1.1754 | -0.7106 |
51
  | 5.00 | 1.3458 | 1.0078 | -1.1730 | -0.7105 |
 
 
 
 
 
 
52
 
53
  Then I followed Rombodawg's [suggestion](https://www.reddit.com/r/LocalLLaMA/comments/1fyx27y/im_pretty_happy_with_how_my_method_worked_out/) to merge [gemma-2-2b](https://huggingface.co/google/gemma-2-2b), [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18) and [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain this model.
54
 
@@ -61,52 +67,14 @@ Click on the model name go to the raw score json generated by Open LLM Leaderboa
61
  | Model | Average | IFEval | BHH | Math Lv5 | GPQA | MUSR | MMLU-PRO |
62
  | ----- | ------- | ------ | ----|--------- | ---- | ---- | -------- |
63
  | [gemma-2-2b-jpn-it](https://huggingface.co/datasets/open-llm-leaderboard/results/blob/main/google/gemma-2-2b-jpn-it/results_2024-10-15T15-21-39.173019.json) | 30.82 | 54.11 | 41.43 | 0.0 | 27.52 | 37.17 | 24.67 |
64
- | gemma-2-2b-ORPO-jpn-it-abliterated-18 (5 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
65
- | gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (5 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
 
 
66
  | [gemma-2-2b-jpn-it-abliterated-17](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-17/results_2024-10-18T15-18-46.821674.json) | 30.29 | 52.65 | 40.46 | 0.0 | 27.18 | 36.90 | 24.55 |
67
  | [gemma-2-2b-jpn-it-abliterated-18](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-18/results_2024-10-18T15-41-42.399571.json) | 30.61 | 53.02 | 40.96 | 0.0 | 27.35 | 37.30 | 25.05 |
68
  | [gemma-2-2b-jpn-it-abliterated-24](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-24/results_2024-10-25T16-29-46.542899.json) | 30.61 | 51.37 | 40.77 | 0.0 | 27.77 | 39.02 | 24.73 |
69
 
70
- ## Merge Details
71
- ### Merge Method
72
-
73
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
74
-
75
- This model was merged using the [TIES](https://arxiv.org/abs/2306.01708) merge method using [google/gemma-2-2b](https://huggingface.co/google/gemma-2-2b) as a base.
76
-
77
- ### Models Merged
78
-
79
- The following models were included in the merge:
80
- * ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
81
- * ymcki/gemma-2-2b-jpn-it-abliterated-18
82
-
83
- ### Configuration
84
-
85
- The following YAML configuration was used to produce this model:
86
-
87
- ```yaml
88
- models:
89
- - model: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
90
- dtype: bfloat16
91
- parameters:
92
- density: 1.0
93
- weight: 1.0
94
- - model: ymcki/gemma-2-2b-jpn-it-abliterated-18
95
- dtype: bfloat16
96
- parameters:
97
- density: 1.0
98
- weight: 1.0
99
- merge_method: ties
100
- base_model: google/gemma-2-2b
101
- parameters:
102
- density: 1.0
103
- weight: 1.0
104
- normalize: true
105
- int8_mask: true
106
- dtype: bfloat16
107
- tokenizer_source: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
108
- ```
109
-
110
  ## How to run this model
111
 
112
  ```py
 
37
 
38
  Using the [gemma-2-2b base model](https://huggingface.co/google/gemma-2-2b), I employed the ORPO method described by [mlabonne](https://towardsdatascience.com/fine-tune-llama-3-with-orpo-56cfab2f9ada) but the input model was read into VRAM by [unsloth](https://github.com/unslothai/unsloth) to allow using the full 40k dataset to run on a single 3090.
39
 
40
+ Five epoches was run. Smallest eval_loss was achieve at epoch 7.72.
41
+ Checkpoint at epoch 7.72 is used to obtain a model adapter and
42
  applied it to [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18).
43
 
44
  | Epoch | loss | eval_loss | eval_logps/rejected | eval_logps/chosen |
 
49
  | 4.00 | 1.5293 | 1.0166 | -1.2004 | -0.7200 |
50
  | 4.96 | 1.2893 | 1.0077 | -1.1754 | -0.7106 |
51
  | 5.00 | 1.3458 | 1.0078 | -1.1730 | -0.7105 |
52
+ | 6.00 | 1.3807 | 0.9924 | -1.1757 | -0.6971 |
53
+ | 7.00 | 1.0855 | 0.9889 | -1.2634 | -0.7235 |
54
+ | 7.72 | 0.8720 | 0.9855 | -1.2374 | -0.7100 |
55
+ | 8.00 | 0.7301 | 0.9864 | -1.2406 | -0.7113 |
56
+ | 9.00 | 1.1939 | 0.9934 | -1.2703 | -0.6852 |
57
+ | 10.00 | 0.7421 | 1.0269 | -1.2552 | -0.7395 |
58
 
59
  Then I followed Rombodawg's [suggestion](https://www.reddit.com/r/LocalLLaMA/comments/1fyx27y/im_pretty_happy_with_how_my_method_worked_out/) to merge [gemma-2-2b](https://huggingface.co/google/gemma-2-2b), [gemma-2-2b-ORPO-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18) and [gemma-2-2b-jpn-it-ablitered-18](https://huggingface.co/ymcki/gemma-2-2b-jpn-it-abliterated-18) to obtain this model.
60
 
 
67
  | Model | Average | IFEval | BHH | Math Lv5 | GPQA | MUSR | MMLU-PRO |
68
  | ----- | ------- | ------ | ----|--------- | ---- | ---- | -------- |
69
  | [gemma-2-2b-jpn-it](https://huggingface.co/datasets/open-llm-leaderboard/results/blob/main/google/gemma-2-2b-jpn-it/results_2024-10-15T15-21-39.173019.json) | 30.82 | 54.11 | 41.43 | 0.0 | 27.52 | 37.17 | 24.67 |
70
+ | [gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (5 epoches)](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18-merge/results_2024-10-30T17-06-58.119904.json) | 29.26 | 49.16 | 38.15 | 2.49 | 28.19 | 33.07 | 24.51 |
71
+ | gemma-2-2b-ORPO-jpn-it-abliterated-18-merge (10 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
72
+ | [gemma-2-2b-ORPO-jpn-it-abliterated-18 (5 epoches)](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18/results_2024-10-30T22-19-29.202883.json) | 29.57 | 48.05 | 41.26 | 0.0 | 27.18 | 36.51 | 24.43
73
+ | gemma-2-2b-ORPO-jpn-it-abliterated-18 (10 epoches) | TBD | TBD | TBD | TBD | TBD | TBD | TBD |
74
  | [gemma-2-2b-jpn-it-abliterated-17](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-17/results_2024-10-18T15-18-46.821674.json) | 30.29 | 52.65 | 40.46 | 0.0 | 27.18 | 36.90 | 24.55 |
75
  | [gemma-2-2b-jpn-it-abliterated-18](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-18/results_2024-10-18T15-41-42.399571.json) | 30.61 | 53.02 | 40.96 | 0.0 | 27.35 | 37.30 | 25.05 |
76
  | [gemma-2-2b-jpn-it-abliterated-24](https://huggingface.co/datasets/open-llm-leaderboard/results/raw/main/ymcki/gemma-2-2b-jpn-it-abliterated-24/results_2024-10-25T16-29-46.542899.json) | 30.61 | 51.37 | 40.77 | 0.0 | 27.77 | 39.02 | 24.73 |
77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
  ## How to run this model
79
 
80
  ```py
mergekit_config.yml CHANGED
@@ -1,10 +1,10 @@
1
  models:
2
- - model: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
3
  dtype: bfloat16
4
  parameters:
5
  density: 1.0
6
  weight: 1.0
7
- - model: ymcki/gemma-2-2b-jpn-it-abliterated-18
8
  dtype: bfloat16
9
  parameters:
10
  density: 1.0
@@ -17,4 +17,4 @@ parameters:
17
  normalize: true
18
  int8_mask: true
19
  dtype: bfloat16
20
- tokenizer_source: ymcki/gemma-2-2b-ORPO-jpn-it-abliterated-18
 
1
  models:
2
+ - model: ./gemma-2-2b-ORPO-jpn-it-abliterated-18
3
  dtype: bfloat16
4
  parameters:
5
  density: 1.0
6
  weight: 1.0
7
+ - model: /home/user/gemma-2-2b-jpn-it-abliterated-18
8
  dtype: bfloat16
9
  parameters:
10
  density: 1.0
 
17
  normalize: true
18
  int8_mask: true
19
  dtype: bfloat16
20
+ tokenizer_source: ./gemma-2-2b-ORPO-jpn-it-abliterated-18
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c19ee69a0923b26a3207aa4153ab24ed61f0a7be73c242b3e9ffbfd03692f2db
3
  size 4959727696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44f62f5fe1d112022beb697189fa7de24d883141a78d8e4cb2c013602c39ac45
3
  size 4959727696
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d385b7de42cc07cf2f39364fa77b4d19a3bfe305750be70113a8d8d0f46d54e0
3
  size 268999016
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c66000edb5b2bd0d28c2aafea08c18cd7324d956e6daca3b5196dc199027492
3
  size 268999016