TheDrummer commited on
Commit
a79ba83
·
verified ·
1 Parent(s): 714f9f6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ tldr; This is Phi 3 Medium finetuned for roleplaying.
2
+
3
+ Training Details:
4
+ - 8x H100 80GB SXM GPUs
5
+ - 10 minutes training duration
6
+ - A continued finetune of Cream-Phi-3-14B-v1b (now released as the official v1)
7
+
8
+ Results for Roleplay Mode (i.e., not Instruct format):
9
+ - Workable RP formatting with occassional mistakes. (Yep, it got worse)
10
+ - Long-ish and moist response. It cooks fast.
11
+ - Slightly incoherent. Can go hard on moist scenes but with poor spatial and anatomical understanding.
12
+ - Important: My testing is lazy and flawed. Take it with a grain of salt and test the GGUFs before taking notes.
13
+
14
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65f2fd1c25b848bd061b5c2e/BjN92w1x9XbsOj0RpALiz.png)
15
+ (No eval split = no eval metrics ^)
16
+
17
+
18
+ Axolotl Config (some fields omitted)
19
+ ```yaml
20
+ base_model: BeaverAI/Cream-Phi-3-14B-v1b
21
+ load_in_4bit: true
22
+ bf16: auto
23
+ fp16:
24
+ tf32: false
25
+ flash_attention: true
26
+
27
+ sequence_len: 6144
28
+ datasets:
29
+ - path: SicariusSicariiStuff/Bluemoon_Top50MB_Sorted_Fixed
30
+ type: customphi3
31
+
32
+ num_epochs: 2
33
+ warmup_steps: 5
34
+ weight_decay: 0.1
35
+
36
+ adapter: lora
37
+ lora_r: 32
38
+ lora_alpha: 16
39
+ lora_dropout: 0.1
40
+ lora_target_linear: true
41
+
42
+ gradient_accumulation_steps: 2
43
+ micro_batch_size: 1
44
+ gradient_checkpointing: true
45
+ gradient_checkpointing_kwargs:
46
+ use_reentrant: true
47
+
48
+ sample_packing: true
49
+ pad_to_sequence_len: true
50
+
51
+ optimizer: paged_adamw_8bit
52
+ lr_scheduler: cosine
53
+ learning_rate: 0.0001
54
+ max_grad_norm: 1.0
55
+ ```