mrfakename commited on
Commit
477575c
1 Parent(s): ce50bb6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -17,7 +17,7 @@ configs:
17
  data_files: zh.json
18
  ---
19
 
20
- Attribution: https://huggingface.co/datasets/wikimedia/wikipedia
21
 
22
  by @mrfakename
23
 
@@ -27,6 +27,22 @@ Only for training StyleTTS 2-related **open source** models.
27
 
28
  processed using: https://huggingface.co/styletts2-community/data-preprocessing-scripts (styletts2 members only)
29
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  ## Note
31
 
32
  East-asian languages are experimental and in beta. We do not distinguish between chinese traditional and simplified, the dataset consists mainly of simplified chinese. We recommend converting characters to simplified chinese during inference using a library such as `hanziconv` or `chinese-converter`.
 
17
  data_files: zh.json
18
  ---
19
 
20
+ #
21
 
22
  by @mrfakename
23
 
 
27
 
28
  processed using: https://huggingface.co/styletts2-community/data-preprocessing-scripts (styletts2 members only)
29
 
30
+ ## License + Credits
31
+
32
+ Source data comes from [Wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) and is licensed under CC-BY-SA 3.0. This dataset is licensed under CC-BY-SA 3.0.
33
+
34
+ ## Processing
35
+
36
+ We utilized the following process to preprocess the dataset:
37
+
38
+ 1. Download data from [Wikipedia](https://huggingface.co/datasets/wikimedia/wikipedia) by language, selecting only the first Parquet file and naming it with the language code
39
+ 2. Process using [https://huggingface.co/styletts2-community/data-preprocessing-scripts](Data Preprocessing Scripts (StyleTTS 2 Community members only)) and modify the code to work with the language
40
+ 3. Clean the text
41
+ 4. Remove ultra-short phrases
42
+ 5. Phonemize
43
+ 6. Save JSON
44
+ 7. Upload dataset
45
+
46
  ## Note
47
 
48
  East-asian languages are experimental and in beta. We do not distinguish between chinese traditional and simplified, the dataset consists mainly of simplified chinese. We recommend converting characters to simplified chinese during inference using a library such as `hanziconv` or `chinese-converter`.