Update README.md
Browse files
README.md
CHANGED
@@ -1,4 +1,6 @@
|
|
1 |
---
|
|
|
|
|
2 |
license: cc-by-4.0
|
3 |
language:
|
4 |
- en
|
@@ -25,16 +27,13 @@ Monologue” method significantly improves the linguistic quality of generated s
|
|
25 |
|
26 |
### Model Sources
|
27 |
|
28 |
-
<!-- Provide the basic links for the model. -->
|
29 |
|
30 |
- **Repository:** [repo](https://github.com/kyutai-labs/moshi)
|
31 |
-
- **Paper:** [paper
|
32 |
- **Demo:** [demo](https://moshi.chat/)
|
33 |
|
34 |
## Uses
|
35 |
|
36 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
37 |
-
|
38 |
### Direct Use
|
39 |
|
40 |
The model can be used as a conversational agent for casual conversations, basic facts and advice (e.g. recipes, trivia), roleplay, etc. However, the model has limited abilities for complex tasks and cannot access tools, but rather focues on natural, low-latency interactions.
|
@@ -54,8 +53,6 @@ This model is for research only and we do not recommend it for providing advices
|
|
54 |
|
55 |
## Bias, Risks, and Limitations
|
56 |
|
57 |
-
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
58 |
-
|
59 |
The model has been trained with a few safeguards to try to limit potential toxic usages, however our toxicity analysis shows that it behaves in the middle of existing models with respect to textual generation. It has some bias towards certain domains and topics that are over-represented in the training data. Its capabilities are relatively limited so far and it is trained to produce only one voice to avoid impersonation. Yet, we need the perspective in time to establish the sociotechnical limitations.
|
60 |
|
61 |
|
@@ -92,16 +89,17 @@ The training was performed on 127 DGX nodes provided by Scaleway, accounting for
|
|
92 |
## Citation
|
93 |
|
94 |
```
|
95 |
-
@
|
96 |
-
|
97 |
-
|
98 |
-
|
99 |
-
|
100 |
-
|
|
|
101 |
}
|
102 |
```
|
103 |
|
104 |
|
105 |
## Model Card Authors
|
106 |
|
107 |
-
Alexandre Défossez, Laurent Mazaré, Manu Orsini, Amélie Royer, Patrick Pérez, Hervé Jégou, Edouard Grave, Neil Zeghidour
|
|
|
1 |
---
|
2 |
+
# For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1
|
3 |
+
# Doc / guide: https://huggingface.co/docs/hub/model-cards
|
4 |
license: cc-by-4.0
|
5 |
language:
|
6 |
- en
|
|
|
27 |
|
28 |
### Model Sources
|
29 |
|
|
|
30 |
|
31 |
- **Repository:** [repo](https://github.com/kyutai-labs/moshi)
|
32 |
+
- **Paper:** [paper](http://kyutai.org/Moshi.pdf)
|
33 |
- **Demo:** [demo](https://moshi.chat/)
|
34 |
|
35 |
## Uses
|
36 |
|
|
|
|
|
37 |
### Direct Use
|
38 |
|
39 |
The model can be used as a conversational agent for casual conversations, basic facts and advice (e.g. recipes, trivia), roleplay, etc. However, the model has limited abilities for complex tasks and cannot access tools, but rather focues on natural, low-latency interactions.
|
|
|
53 |
|
54 |
## Bias, Risks, and Limitations
|
55 |
|
|
|
|
|
56 |
The model has been trained with a few safeguards to try to limit potential toxic usages, however our toxicity analysis shows that it behaves in the middle of existing models with respect to textual generation. It has some bias towards certain domains and topics that are over-represented in the training data. Its capabilities are relatively limited so far and it is trained to produce only one voice to avoid impersonation. Yet, we need the perspective in time to establish the sociotechnical limitations.
|
57 |
|
58 |
|
|
|
89 |
## Citation
|
90 |
|
91 |
```
|
92 |
+
@techreport{citation-key,
|
93 |
+
author = {Alexandre D\'efossez, Laurent Mazar\'e, Manu Orsini, Am\'elie Royer, Patrick P\'erez, Herv\'e J\'egou, Edouard Grave, Neil Zeghidour},
|
94 |
+
title = {Moshi: a speech-text foundation model for real-time dialogue},
|
95 |
+
institution = {Kyutai},
|
96 |
+
year={2024},
|
97 |
+
month={September},
|
98 |
+
url={http://kyutai.org/Moshi.pdf},
|
99 |
}
|
100 |
```
|
101 |
|
102 |
|
103 |
## Model Card Authors
|
104 |
|
105 |
+
Alexandre Défossez, Laurent Mazaré, Manu Orsini, Amélie Royer, Patrick Pérez, Hervé Jégou, Edouard Grave, Neil Zeghidour
|