pyf98 commited on
Commit
582bb63
1 Parent(s): d49740e

update text

Browse files
Files changed (1) hide show
  1. app.py +32 -5
app.py CHANGED
@@ -23,19 +23,46 @@ OWSM v3.1 has 1.02B parameters and is trained on 180k hours of paired speech dat
23
  - Long-form transcription
24
  - Language identification
25
 
 
 
 
 
 
 
 
 
 
 
26
  ```
27
- @article{peng2023owsm,
28
  title={Reproducing Whisper-Style Training Using an Open-Source Toolkit and Publicly Available Data},
29
  author={Yifan Peng and Jinchuan Tian and Brian Yan and Dan Berrebbi and Xuankai Chang and Xinjian Li and Jiatong Shi and Siddhant Arora and William Chen and Roshan Sharma and Wangyou Zhang and Yui Sudo and Muhammad Shakeel and Jee-weon Jung and Soumi Maiti and Shinji Watanabe},
30
- journal={arXiv preprint arXiv:2309.13876},
31
  year={2023}
32
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
33
  ```
34
 
35
- As a demo, the input speech should not exceed 2 minutes. We also limit the maximum number of tokens to be generated.
36
- Please try our [Colab demo](https://colab.research.google.com/drive/1zKI3ZY_OtZd6YmVeED6Cxy1QwT1mqv9O?usp=sharing) if you want to explore more features.
37
 
38
- Disclaimer: OWSM has not been thoroughly evaluated in all tasks. Due to limited training data, it may not perform well for certain language directions.
39
  '''
40
 
41
  if not torch.cuda.is_available():
 
23
  - Long-form transcription
24
  - Language identification
25
 
26
+ As a demo, the input speech should not exceed 2 minutes. We also limit the maximum number of tokens to be generated.
27
+ Please try our [Colab demo](https://colab.research.google.com/drive/1zKI3ZY_OtZd6YmVeED6Cxy1QwT1mqv9O?usp=sharing) if you want to explore more features.
28
+
29
+ Disclaimer: OWSM has not been thoroughly evaluated in all tasks. Due to limited training data, it may not perform well for certain language directions.
30
+
31
+ Please consider citing the following related papers if you find our work helpful.
32
+
33
+ <details><summary>citations</summary>
34
+ <p>
35
+
36
  ```
37
+ @inproceedings{peng2023owsm,
38
  title={Reproducing Whisper-Style Training Using an Open-Source Toolkit and Publicly Available Data},
39
  author={Yifan Peng and Jinchuan Tian and Brian Yan and Dan Berrebbi and Xuankai Chang and Xinjian Li and Jiatong Shi and Siddhant Arora and William Chen and Roshan Sharma and Wangyou Zhang and Yui Sudo and Muhammad Shakeel and Jee-weon Jung and Soumi Maiti and Shinji Watanabe},
40
+ booktitle={Proc. ASRU},
41
  year={2023}
42
  }
43
+ @inproceedings{peng23b_interspeech,
44
+ author={Yifan Peng and Kwangyoun Kim and Felix Wu and Brian Yan and Siddhant Arora and William Chen and Jiyang Tang and Suwon Shon and Prashant Sridhar and Shinji Watanabe},
45
+ title={{A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks}},
46
+ year=2023,
47
+ booktitle={Proc. INTERSPEECH},
48
+ }
49
+ @inproceedings{kim2023branchformer,
50
+ title={E-branchformer: Branchformer with enhanced merging for speech recognition},
51
+ author={Kim, Kwangyoun and Wu, Felix and Peng, Yifan and Pan, Jing and Sridhar, Prashant and Han, Kyu J and Watanabe, Shinji},
52
+ booktitle={2022 IEEE Spoken Language Technology Workshop (SLT)},
53
+ year={2023},
54
+ }
55
+ @InProceedings{pmlr-v162-peng22a,
56
+ title = {Branchformer: Parallel {MLP}-Attention Architectures to Capture Local and Global Context for Speech Recognition and Understanding},
57
+ author = {Peng, Yifan and Dalmia, Siddharth and Lane, Ian and Watanabe, Shinji},
58
+ booktitle = {Proceedings of the 39th International Conference on Machine Learning},
59
+ year = {2022},
60
+ }
61
  ```
62
 
63
+ </p>
64
+ </details>
65
 
 
66
  '''
67
 
68
  if not torch.cuda.is_available():