Akito-UzukiP commited on
Commit
d1a5f75
·
1 Parent(s): 77d2471

update models

Browse files
app.py CHANGED
@@ -123,7 +123,7 @@ def tts_fn(text, speaker, sdp_ratio, noise_scale, noise_scale_w, length_scale, l
123
  if __name__ == "__main__":
124
  parser = argparse.ArgumentParser()
125
  parser.add_argument(
126
- "-m", "--model", default="./logs/umamusume/G_138000.pth", help="path of your model"
127
  )
128
  parser.add_argument(
129
  "-c",
@@ -161,6 +161,7 @@ if __name__ == "__main__":
161
  **hps.model,
162
  ).to(device)
163
  _ = net_g.eval()
 
164
 
165
  _ = utils.load_checkpoint(args.model, net_g, None, skip_optimizer=True)
166
 
 
123
  if __name__ == "__main__":
124
  parser = argparse.ArgumentParser()
125
  parser.add_argument(
126
+ "-m", "--model", default="./logs/umamusume/G_211000.pth", help="path of your model"
127
  )
128
  parser.add_argument(
129
  "-c",
 
161
  **hps.model,
162
  ).to(device)
163
  _ = net_g.eval()
164
+ _ = net_g.half()
165
 
166
  _ = utils.load_checkpoint(args.model, net_g, None, skip_optimizer=True)
167
 
logs/umamusume/D_138000.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:468ec1276c5524d9ea67149eb7c3867212d01976a56373a03668c97b8f6fab67
3
- size 561070759
 
 
 
 
logs/umamusume/G_138000.pth DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:93a7f35fa578f375c48cc5a72166c9a54cb972b69778c35cb601c814990394b3
3
- size 857607936
 
 
 
 
logs/umamusume/{DUR_138000.pth → G_211000.pth} RENAMED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:895126bae99723209956cfd0cae65c33899e6a3f61f93ce3346818876f1dbe69
3
- size 6885803
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68c344dd054d8777e33aae66cf351e5012d0aa00a1602a68900623375d4a0c03
3
+ size 857605376
text/japanese.py CHANGED
@@ -658,11 +658,11 @@ import os
658
  if __name__ == "__main__":
659
  tokenizer = AutoTokenizer.from_pretrained("cl-tohoku/bert-base-japanese-v3")
660
  #tokenizer = AutoTokenizer.from_pretrained("bert/bert-base-japanese-v3")
661
- text = "これが先頭の景色……観覧車みたいです。童、小童!"
662
  from text.japanese_bert import get_bert_feature
663
 
664
-
665
  phones, tones, word2ph = g2p(text)
666
  bert = get_bert_feature(text, word2ph)
667
-
668
- print(phones, tones, word2ph,bert.shape)
 
 
658
  if __name__ == "__main__":
659
  tokenizer = AutoTokenizer.from_pretrained("cl-tohoku/bert-base-japanese-v3")
660
  #tokenizer = AutoTokenizer.from_pretrained("bert/bert-base-japanese-v3")
661
+ text = "これが先頭の景色……観覧車みたい、美味しそうな弥生賞!貴方は一人ですか。童、小童!"
662
  from text.japanese_bert import get_bert_feature
663
 
 
664
  phones, tones, word2ph = g2p(text)
665
  bert = get_bert_feature(text, word2ph)
666
+ tagger = MeCab.Tagger()
667
+ print(tagger.parse(text))
668
+ # print(phones, tones, word2ph,bert.shape)