mikeforai commited on
Commit
1cd75d2
·
verified ·
1 Parent(s): fe9230f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -2
README.md CHANGED
@@ -2,8 +2,7 @@
2
  license: cc-by-nc-4.0
3
  pipeline_tag: text-to-audio
4
  ---
5
- Model from facebook/musicgen-medium
6
 
7
- MusicGen - Medium - 1.5B
8
 
9
  MusicGen is a text-to-music model capable of genreating high-quality music samples conditioned on text descriptions or audio prompts. It is a single stage auto-regressive Transformer model trained over a 32kHz EnCodec tokenizer with 4 codebooks sampled at 50 Hz. Unlike existing methods, like MusicLM, MusicGen doesn't require a self-supervised semantic representation, and it generates all 4 codebooks in one pass. By introducing a small delay between the codebooks, we show we can predict them in parallel, thus having only 50 auto-regressive steps per second of audio.
 
2
  license: cc-by-nc-4.0
3
  pipeline_tag: text-to-audio
4
  ---
 
5
 
6
+ # This is the place where converting facebook/musicgen-medium safetensors to PyTorch .bin Model for easier usage for everyone
7
 
8
  MusicGen is a text-to-music model capable of genreating high-quality music samples conditioned on text descriptions or audio prompts. It is a single stage auto-regressive Transformer model trained over a 32kHz EnCodec tokenizer with 4 codebooks sampled at 50 Hz. Unlike existing methods, like MusicLM, MusicGen doesn't require a self-supervised semantic representation, and it generates all 4 codebooks in one pass. By introducing a small delay between the codebooks, we show we can predict them in parallel, thus having only 50 auto-regressive steps per second of audio.