NeMo
CasanovaE commited on
Commit
c7f79fa
1 Parent(s): e00df4b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +6 -2
README.md CHANGED
@@ -23,9 +23,13 @@ The [Low Frame-rate Speech Codec](https://arxiv.org/abs/2409.12117) is a neural
23
 
24
 
25
  ## Model Architecture
26
- Low Frame-rate Speech Codec model is composed of a fully convolutional generator neural network and three discriminators. The generator comprises an encoder, followed by vector quantization, and a [HiFi-GAN-based](https://arxiv.org/abs/2010.05646) decoder.
27
- For the discriminators, we utilize three neural networks, all employing a squared-GAN and feature-matching loss. We adopt the multi-period discriminator proposed by \cite{kong2020hifi} and the multi-scale complex STFT discriminator proposed by \cite{defossez2022high}. Additionally, inspired by \cite{li2024styletts}, we proposed the use of Speech Language Models (SLMs) as a discriminator. SLMs encode information ranging from acoustic to semantic aspects, which could benefit our model's training, especially in low frame rate settings where accurate pronunciation is difficult to achieve due to the high compression rate. We adopted the 12-layer WavLM \cite{chen2022wavlm}, pre-trained on 94k hours of data, as the SLM. During training, we resample the input audio to 16 kHz before feeding it into the WavLM model, extracting the intermediary layer features. These features are then fed to a discriminative head composed of four 1D convolutional layers. As in \cite{chen2022wavlm}, the SLM remains frozen during training.
 
 
 
28
 
 
29
 
30
  ### Input
31
  - **Input Type:** Audio
 
23
 
24
 
25
  ## Model Architecture
26
+ Low Frame-rate Speech Codec model is composed of a fully convolutional generator neural network and three discriminators.
27
+ The generator comprises an encoder, followed by vector quantization, and a [HiFi-GAN-based](https://arxiv.org/abs/2010.05646) decoder.
28
+ The encoder consists of five residual blocks, each block containing three residual layers similar to the [multi-receptive field fusion (MRF) module](https://arxiv.org/abs/2010.05646). For the vector quantization, we have used [Finite Scalar Quantization (FSQ)](https://arxiv.org/abs/2309.15505) with eight codebooks and four dimensions per code and 2016 codes per codebook.
29
+ For the discriminators, we utilize three neural networks, all employing a squared-GAN and feature-matching loss. We adopt the [multi-period discriminator](https://arxiv.org/abs/2010.05646) and the [multi-scale complex STFT discriminator](https://arxiv.org/abs/2210.13438).
30
+ Additionally, we proposed the use of Speech Language Models (SLMs) as a discriminator. SLMs encode information ranging from acoustic to semantic aspects, which could benefit our model's training, especially in low frame rate settings where accurate pronunciation is difficult to achieve due to the high compression rate. We adopted the [12-layer WavLM](https://arxiv.org/abs/2110.13900), pre-trained on 94k hours of data, as the SLM. During training, we resample the input audio to 16 kHz before feeding it into the WavLM model, extracting the intermediary layer features. These features are then fed to a discriminative head composed of four 1D convolutional layers.
31
 
32
+ For more details please check [our paper](https://arxiv.org/abs/2409.12117).
33
 
34
  ### Input
35
  - **Input Type:** Audio