machine-teaching-group commited on
Commit
9b40a44
·
verified ·
1 Parent(s): 47050c4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +51 -3
README.md CHANGED
@@ -1,3 +1,51 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ base_model:
4
+ - course-genai-w24/week4-phi-1.5-sft-shakespeare
5
+ ---
6
+ # Model Card for Model ID
7
+
8
+ ### Summary
9
+
10
+ <!-- Provide a quick summary of what the model is/does. -->
11
+
12
+ This is a preference tuned model for text completion based on Phi 1.5. It has been tuned on a filtered version of the The Complete Works of William Shakespeare, which can be found and downloaded from here: [https://www.gutenberg.org/ebooks/100](https://www.gutenberg.org/ebooks/100).
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+ - **Developed by:** Course Organizers
19
+ - **Finetuned from model:** microsoft/phi-1_5
20
+
21
+ ### Training Details
22
+
23
+ This model has been trained using the TRL library and OPROTrainer class from Huggingface.
24
+
25
+ ### Training Data
26
+
27
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
28
+
29
+ The Complete Works of William Shakespeare, which can be found and downloaded from here: [https://www.gutenberg.org/ebooks/100](https://www.gutenberg.org/ebooks/100)
30
+
31
+ #### Training Hyperparameters
32
+
33
+ The following hyperparameters were used during training:
34
+
35
+ - learning_rate: 1e-06
36
+ - per_device_train_batch_size: 1
37
+ - lr_scheduler_type: cosine
38
+ - weight_decay: 0.01
39
+ - num_epochs: 1
40
+
41
+
42
+ ### Framework Versions
43
+
44
+ - accelerate==0.26.1
45
+ - datasets==2.16.1
46
+ - transformers==4.45.2
47
+ - trl==0.11.2
48
+
49
+ ### Compute Infrastructure and Hardware
50
+
51
+ Slurm cluster with 8 x H100 Nvidia GPUs.