Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,51 @@
|
|
1 |
-
---
|
2 |
-
license: mit
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: mit
|
3 |
+
base_model:
|
4 |
+
- course-genai-w24/week4-phi-1.5-sft-shakespeare
|
5 |
+
---
|
6 |
+
# Model Card for Model ID
|
7 |
+
|
8 |
+
### Summary
|
9 |
+
|
10 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
+
|
12 |
+
This is a preference tuned model for text completion based on Phi 1.5. It has been tuned on a filtered version of the The Complete Works of William Shakespeare, which can be found and downloaded from here: [https://www.gutenberg.org/ebooks/100](https://www.gutenberg.org/ebooks/100).
|
13 |
+
|
14 |
+
### Model Description
|
15 |
+
|
16 |
+
<!-- Provide a longer summary of what this model is. -->
|
17 |
+
|
18 |
+
- **Developed by:** Course Organizers
|
19 |
+
- **Finetuned from model:** microsoft/phi-1_5
|
20 |
+
|
21 |
+
### Training Details
|
22 |
+
|
23 |
+
This model has been trained using the TRL library and OPROTrainer class from Huggingface.
|
24 |
+
|
25 |
+
### Training Data
|
26 |
+
|
27 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
28 |
+
|
29 |
+
The Complete Works of William Shakespeare, which can be found and downloaded from here: [https://www.gutenberg.org/ebooks/100](https://www.gutenberg.org/ebooks/100)
|
30 |
+
|
31 |
+
#### Training Hyperparameters
|
32 |
+
|
33 |
+
The following hyperparameters were used during training:
|
34 |
+
|
35 |
+
- learning_rate: 1e-06
|
36 |
+
- per_device_train_batch_size: 1
|
37 |
+
- lr_scheduler_type: cosine
|
38 |
+
- weight_decay: 0.01
|
39 |
+
- num_epochs: 1
|
40 |
+
|
41 |
+
|
42 |
+
### Framework Versions
|
43 |
+
|
44 |
+
- accelerate==0.26.1
|
45 |
+
- datasets==2.16.1
|
46 |
+
- transformers==4.45.2
|
47 |
+
- trl==0.11.2
|
48 |
+
|
49 |
+
### Compute Infrastructure and Hardware
|
50 |
+
|
51 |
+
Slurm cluster with 8 x H100 Nvidia GPUs.
|