jegou commited on
Commit
f45037c
·
verified ·
1 Parent(s): 894dc28

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -13
README.md CHANGED
@@ -1,4 +1,6 @@
1
  ---
 
 
2
  license: cc-by-4.0
3
  language:
4
  - en
@@ -25,16 +27,13 @@ Monologue” method significantly improves the linguistic quality of generated s
25
 
26
  ### Model Sources
27
 
28
- <!-- Provide the basic links for the model. -->
29
 
30
  - **Repository:** [repo](https://github.com/kyutai-labs/moshi)
31
- - **Paper:** [paper (soon!)](https://arxiv.org/abs/2409.XXXXX)
32
  - **Demo:** [demo](https://moshi.chat/)
33
 
34
  ## Uses
35
 
36
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
37
-
38
  ### Direct Use
39
 
40
  The model can be used as a conversational agent for casual conversations, basic facts and advice (e.g. recipes, trivia), roleplay, etc. However, the model has limited abilities for complex tasks and cannot access tools, but rather focues on natural, low-latency interactions.
@@ -54,8 +53,6 @@ This model is for research only and we do not recommend it for providing advices
54
 
55
  ## Bias, Risks, and Limitations
56
 
57
- <!-- This section is meant to convey both technical and sociotechnical limitations. -->
58
-
59
  The model has been trained with a few safeguards to try to limit potential toxic usages, however our toxicity analysis shows that it behaves in the middle of existing models with respect to textual generation. It has some bias towards certain domains and topics that are over-represented in the training data. Its capabilities are relatively limited so far and it is trained to produce only one voice to avoid impersonation. Yet, we need the perspective in time to establish the sociotechnical limitations.
60
 
61
 
@@ -92,16 +89,17 @@ The training was performed on 127 DGX nodes provided by Scaleway, accounting for
92
  ## Citation
93
 
94
  ```
95
- @article{defossez2024moshi,
96
- title={Moshi: a speech-text foundation model for real-time dialogue},
97
- authors={Alexandre D\'efossez, Laurent Mazar\'e, Manu Orsini, Am\'elie Royer, Patrick P\'erez, Herv\'e J\'egou, Edouard Grave, Neil Zeghidour},
98
- year={2024},
99
- month={September},
100
- journal={arXiv preprint arXiv:2409.XXXXX}
 
101
  }
102
  ```
103
 
104
 
105
  ## Model Card Authors
106
 
107
- Alexandre Défossez, Laurent Mazaré, Manu Orsini, Amélie Royer, Patrick Pérez, Hervé Jégou, Edouard Grave, Neil Zeghidour
 
1
  ---
2
+ # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
3
+ # Doc / guide: https://huggingface.co/docs/hub/model-cards
4
  license: cc-by-4.0
5
  language:
6
  - en
 
27
 
28
  ### Model Sources
29
 
 
30
 
31
  - **Repository:** [repo](https://github.com/kyutai-labs/moshi)
32
+ - **Paper:** [paper](http://kyutai.org/Moshi.pdf)
33
  - **Demo:** [demo](https://moshi.chat/)
34
 
35
  ## Uses
36
 
 
 
37
  ### Direct Use
38
 
39
  The model can be used as a conversational agent for casual conversations, basic facts and advice (e.g. recipes, trivia), roleplay, etc. However, the model has limited abilities for complex tasks and cannot access tools, but rather focues on natural, low-latency interactions.
 
53
 
54
  ## Bias, Risks, and Limitations
55
 
 
 
56
  The model has been trained with a few safeguards to try to limit potential toxic usages, however our toxicity analysis shows that it behaves in the middle of existing models with respect to textual generation. It has some bias towards certain domains and topics that are over-represented in the training data. Its capabilities are relatively limited so far and it is trained to produce only one voice to avoid impersonation. Yet, we need the perspective in time to establish the sociotechnical limitations.
57
 
58
 
 
89
  ## Citation
90
 
91
  ```
92
+ @techreport{citation-key,
93
+ author = {Alexandre D\'efossez, Laurent Mazar\'e, Manu Orsini, Am\'elie Royer, Patrick P\'erez, Herv\'e J\'egou, Edouard Grave, Neil Zeghidour},
94
+ title = {Moshi: a speech-text foundation model for real-time dialogue},
95
+ institution = {Kyutai},
96
+ year={2024},
97
+ month={September},
98
+ url={http://kyutai.org/Moshi.pdf},
99
  }
100
  ```
101
 
102
 
103
  ## Model Card Authors
104
 
105
+ Alexandre Défossez, Laurent Mazaré, Manu Orsini, Amélie Royer, Patrick Pérez, Hervé Jégou, Edouard Grave, Neil Zeghidour