Spaces:
Sleeping
Sleeping
Update README.md
Browse files
README.md
CHANGED
@@ -95,7 +95,7 @@ Using the data with additional tokens, I generated two datasets for Rachel in En
|
|
95 |

|
96 |
|
97 |
# Training
|
98 |
-
For training the model to transfer style of Rachel to chatbot I used several models.
|
99 |
1. First one is GPT2. For datasets I used TextDataset by PyTorch and library transformers by huggingface.
|
100 |
The results are shown on the image below
|
101 |

|
@@ -113,4 +113,16 @@ For training the model to transfer style of Rachel to chatbot I used several mod
|
|
113 |

|
114 |
Result of the training is shown below
|
115 |

|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
116 |
|
|
|
95 |

|
96 |
|
97 |
# Training
|
98 |
+
For training the model to transfer style of Rachel to chatbot I used several models. Models are trained in two stages. First stage, when the model tries to catch the personal of Rachel and learns her monologues. Second step the model tries to learn how the Rachel behaves in dialoges, so at this stage model is trained on dialogues.
|
99 |
1. First one is GPT2. For datasets I used TextDataset by PyTorch and library transformers by huggingface.
|
100 |
The results are shown on the image below
|
101 |

|
|
|
113 |

|
114 |
Result of the training is shown below
|
115 |

|
116 |
+
|
117 |
+
# Architecture
|
118 |
+
|
119 |
+
- PrepareData.ipynb <- Parser data from the web, clean, tokenize and prepare for datasets
|
120 |
+
- train_data <- datasets folder with monologues and dialogs
|
121 |
+
- Training_gpt2_medium.ipynb <- training gpt2-medium
|
122 |
+
- en_gpt2-medium_rachel_replics <- gpt2-medium model
|
123 |
+
- Training_gpt2_large.ipynb <- training gpt2-large
|
124 |
+
- en_gpt2-large_rachel_replics <- gpt2-large model
|
125 |
+
- images <- images for README.md
|
126 |
+
- app.py <- main file
|
127 |
+
- requirements.txt <- needed libraries
|
128 |
|