Text Generation
Transformers
Safetensors
English
gemma2
creative-writing
conversational
text-generation-inference
Inference Endpoints
mehmetkeremturkcan commited on
Commit
a8e0ed5
·
verified ·
1 Parent(s): 15bdcfa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -3
README.md CHANGED
@@ -1,3 +1,55 @@
1
- ---
2
- license: gemma
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: gemma
3
+ datasets:
4
+ - sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo
5
+ - sam-paech/gutenbergs_1_2_3_antislop-dpo
6
+ language:
7
+ - en
8
+ base_model:
9
+ - google/gemma-2-9b-it
10
+ library_name: transformers
11
+ tags:
12
+ - creative-writing
13
+ ---
14
+ # Oblivion's End
15
+
16
+ A merged LoRA for gemma-2-9b-it, trained using DPO datasets for creative writing using [my DPO training notebook](https://github.com/mkturkcan/dpo-model-trainer).
17
+
18
+ ## Model Details
19
+
20
+ ### Model Description
21
+
22
+
23
+ - **Finetuned from model:** google/gemma-2-9b-it
24
+ ### Model Sources [optional]
25
+
26
+ - **Repository:** [GitHub](https://github.com/mkturkcan/dpo-model-trainer/tree/main).
27
+
28
+ ## Uses
29
+
30
+ Made for creative writing.
31
+
32
+ ## Training Details
33
+
34
+ ### Training Data
35
+
36
+ Check out the model card details.
37
+
38
+ ### Training Procedure
39
+
40
+ Model training performance (margins) are available in the [wandb instance](https://api.wandb.ai/links/mkturkcan/4djkmhwp).
41
+
42
+ #### Training Hyperparameters
43
+
44
+ - **Training regime:** bf16 on a 1x 80GB A100 node.
45
+
46
+ ## Model Examination [optional]
47
+
48
+ <!-- Relevant interpretability work for the model goes here -->
49
+
50
+ [More Information Needed]
51
+
52
+ ## Environmental Impact
53
+
54
+ Total emissions are estimated to be 0.83 kgCO$_2$eq.
55
+