Xin Xu
commited on
Commit
·
172276e
1
Parent(s):
c3b7826
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,44 @@
|
|
1 |
---
|
2 |
license: apache-2.0
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: apache-2.0
|
3 |
+
language:
|
4 |
+
- en
|
5 |
+
tags:
|
6 |
+
- MuseCoco
|
7 |
+
- Text2Music
|
8 |
---
|
9 |
+
|
10 |
+
<p align="center" width="100%">
|
11 |
+
<a href="" target="_blank"><img src="https://ai-muzic.github.io/images/musecoco/framework.png" alt="Text2Attribute" style="width: 60%; min-width: 60px; display: block; margin: auto;"></a>
|
12 |
+
</p>
|
13 |
+
|
14 |
+
# Text-to-Attribute Understanding
|
15 |
+
This is the text-to-attribute model to extract musical attributes from text, introduced in the paper [*MuseCoco: Generating Symbolic Music from Text*](https://arxiv.org/abs/2306.00110) and [first released in this repository](https://github.com/microsoft/muzic/tree/main/musecoco).
|
16 |
+
It is based on BERT-large and has multiple classification heads for diverse musical attributes.
|
17 |
+
|
18 |
+
There is the mapping between keywords used in the model and musical attributes:
|
19 |
+
```json
|
20 |
+
{
|
21 |
+
"I1s2": "Instrument",
|
22 |
+
"R1": "Rhythm Danceability",
|
23 |
+
"R3": "Rhythm Intensity",
|
24 |
+
"S2s1": "Artist",
|
25 |
+
"S4": "Genre",
|
26 |
+
"B1s1": "Bar",
|
27 |
+
"TS1s1": "Time Signature",
|
28 |
+
"K1": "Key",
|
29 |
+
"T1s1": "Tempo",
|
30 |
+
"P4": "Pitch Range",
|
31 |
+
"EM1": "Emotion",
|
32 |
+
"TM1": "Time"
|
33 |
+
}
|
34 |
+
```
|
35 |
+
|
36 |
+
# BibTeX entry and citation info
|
37 |
+
```bibtex
|
38 |
+
@article{musecoco2023,
|
39 |
+
title={MuseCoco: Generating Symbolic Music from Text},
|
40 |
+
author={Peiling Lu, Xin Xu, Chenfei Kang, Botao Yu, Chengyi Xing, Xu Tan, Jiang Bian},
|
41 |
+
journal={arXiv preprint arXiv:2306.00110},
|
42 |
+
year={2023}
|
43 |
+
}
|
44 |
+
```
|