georgesung commited on
Commit
7c50e44
1 Parent(s): 586386b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -0
README.md CHANGED
@@ -8,6 +8,23 @@ datasets:
8
  Fine-tuned [OpenLLaMA-7B](https://huggingface.co/openlm-research/open_llama_7b) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
9
  Used QLoRA for fine-tuning. Trained for one epoch on a 24GB GPU (NVIDIA A10G) instance, took ~18 hours to train.
10
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  # Training code
12
  Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
13
 
 
8
  Fine-tuned [OpenLLaMA-7B](https://huggingface.co/openlm-research/open_llama_7b) with an uncensored/unfiltered Wizard-Vicuna conversation dataset [ehartford/wizard_vicuna_70k_unfiltered](https://huggingface.co/datasets/ehartford/wizard_vicuna_70k_unfiltered).
9
  Used QLoRA for fine-tuning. Trained for one epoch on a 24GB GPU (NVIDIA A10G) instance, took ~18 hours to train.
10
 
11
+ # Prompt style
12
+ The model was trained with the following prompt style:
13
+ ```
14
+ ### HUMAN:
15
+ Hello
16
+
17
+ ### RESPONSE:
18
+ Hi, how are you?
19
+
20
+ ### HUMAN:
21
+ I'm fine.
22
+
23
+ ### RESPONSE:
24
+ How can I help you?
25
+ ...
26
+ ```
27
+
28
  # Training code
29
  Code used to train the model is available [here](https://github.com/georgesung/llm_qlora).
30