krum-utsav commited on
Commit
448e359
·
1 Parent(s): d2accc0

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +46 -1
README.md CHANGED
@@ -1,6 +1,51 @@
1
  ---
2
  library_name: peft
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  ## Training procedure
5
 
6
 
@@ -17,4 +62,4 @@ The following `bitsandbytes` quantization config was used during training:
17
  ### Framework versions
18
 
19
 
20
- - PEFT 0.4.0.dev0
 
1
  ---
2
  library_name: peft
3
+ license: wtfpl
4
+ language:
5
+ - en
6
+ pipeline_tag: text-generation
7
  ---
8
+
9
+ ## Model description
10
+
11
+ togethercomputer/RedPajama-INCITE-Base-3B-v1 finetuned for paraphrasing and changing the tone of the input sentence(to casual/professional/witty).
12
+ Sample training data:
13
+ ```json
14
+ {
15
+ "original": "If you have any further questions, feel free to ask.",
16
+ "casual": "Got more questions? Feel free to ask away. I'm here to help!",
17
+ "professional": "Should you have any additional inquiries, please don't hesitate to ask.",
18
+ "witty": "Curiosity is always in style! If you have more mysteries to solve, I'm all ears!",
19
+ "paraphrase": "Don't hesitate to ask if you have any more questions."
20
+ }
21
+ ```
22
+
23
+ ## Training params
24
+
25
+ ```json
26
+ {
27
+ "batch_size": 8,
28
+ "eval_ratio": 0.1,
29
+ "eval_steps": 100,
30
+ "gradient_accumulation_steps": 1,
31
+ "learning_rate": 0.0001,
32
+ "logging_steps": 100,
33
+ "lora_alpha": 32,
34
+ "lora_dropout": 0.05,
35
+ "lora_r": 16,
36
+ "max_length": 128,
37
+ "model_name": "togethercomputer/RedPajama-INCITE-Base-3B-v1",
38
+ "num_train_epochs": 3,
39
+ "seed": 10,
40
+ "task_type": "paraphrase_tone",
41
+ "use_aim": True
42
+ }
43
+ ```
44
+
45
+ ## Training curve
46
+
47
+ ![train_eval_loss](RedPajama-INCITE-Base-3B-v1-paraphrase-tone.jpeg)
48
+
49
  ## Training procedure
50
 
51
 
 
62
  ### Framework versions
63
 
64
 
65
+ - PEFT 0.4.0.dev0