morenolq commited on
Commit
3354fe9
·
verified ·
1 Parent(s): 5a15d6b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +38 -8
README.md CHANGED
@@ -1,26 +1,56 @@
1
  ---
2
  license: mit
 
 
 
 
3
  ---
4
 
5
- # SSL4PR WavLM Base
6
 
7
- This repository hosts the pre-trained SSL4PR WavLM Base models for Parkinson's Disease detection from speech in real-world operating conditions. These models are based on the work titled "Exploiting Foundation Models and Speech Enhancement for Parkinsons Disease Detection from Speech in Real-World Operative Conditions" by Moreno La Quatra et al.
8
 
9
  ## Repository Link
10
- [GitHub Repository](https://github.com/K-STMLab/SSL4PR/)
 
11
 
12
  ## Pre-trained Models
13
- Pre-trained models are available on the Hugging Face model hub. To use the SSL4PR WavLM Base models, please clone the repository by running the following command:
 
14
 
15
  ```bash
 
 
16
  git clone https://huggingface.co/morenolq/SSL4PR-hubert-base
 
 
 
 
17
  ```
18
 
19
- Ensure you have git lfs installed. Each repository contains the pre-trained models, one per fold, named `fold_1.pt`, `fold_2.pt`, ..., `fold_10.pt`.
20
- The models are available in PyTorch format.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
21
 
22
- - [SSL4PR WavLM Base](https://huggingface.co/morenolq/SSL4PR-wavlm-base)
23
- - [SSL4PR HuBERT Base](https://huggingface.co/morenolq/SSL4PR-hubert-base) - **this repository**
24
 
25
  ## Citation
26
 
 
1
  ---
2
  license: mit
3
+ tags:
4
+ - dysarthric speech
5
+ - classification
6
+ - audio classification
7
  ---
8
 
9
+ # SSL4PR WavLM Base and HuBERT Base Models
10
 
11
+ This repository hosts the pre-trained SSL4PR models for Parkinson's Disease detection from speech in real-world operating conditions. These models are based on the work titled "Exploiting Foundation Models and Speech Enhancement for Parkinson's Disease Detection from Speech in Real-World Operative Conditions" by La Quatra et al. published at Interspeech 2024. [Paper Link](https://www.isca-archive.org/interspeech_2024/laquatra24_interspeech.pdf)
12
 
13
  ## Repository Link
14
+
15
+ [GitHub Repository](https://github.com/K-STMLab/SSL4PR/) please refer to the repository for all details on the models, training and usage.
16
 
17
  ## Pre-trained Models
18
+
19
+ Pre-trained models are available on the Hugging Face model hub. To use the SSL4PR models, please clone the desired repository by running one of the following commands:
20
 
21
  ```bash
22
+ # For fold-based models (10-fold cross-validation)
23
+ git clone https://huggingface.co/morenolq/SSL4PR-wavlm-base
24
  git clone https://huggingface.co/morenolq/SSL4PR-hubert-base
25
+
26
+ # For full training models (trained on complete s-PC-GITA)
27
+ git clone https://huggingface.co/morenolq/SSL4PR-wavlm-base-full
28
+ git clone https://huggingface.co/morenolq/SSL4PR-hubert-base-full
29
  ```
30
 
31
+ Ensure you have git lfs installed.
32
+
33
+ ### Fold-based Models
34
+ The fold-based repositories contain models trained using 10-fold cross-validation on s-PC-GITA.
35
+ Each repository contains 10 pre-trained models, one per fold, named `fold_1.pt`, `fold_2.pt`, ..., `fold_10.pt`.
36
+
37
+ - [SSL4PR WavLM Base](https://huggingface.co/morenolq/SSL4PR-wavlm-base): using as base model [WavLM Base](https://huggingface.co/microsoft/wavlm-base)
38
+ - [SSL4PR HuBERT Base](https://huggingface.co/morenolq/SSL4PR-hubert-base): using as base model [HuBERT Base](https://huggingface.co/facebook/hubert-base-ls960)
39
+
40
+ ### Full Training Models
41
+ The full training repositories contain models trained on the complete s-PC-GITA dataset and tested on
42
+ enhanced e-PC-GITA (as reported in Table 3 of the paper). Each repository contains a single model file
43
+ named `model.pt`.
44
+
45
+ - [SSL4PR WavLM Base Full](https://huggingface.co/morenolq/SSL4PR-wavlm-base-full): using as base model [WavLM Base](https://huggingface.co/microsoft/wavlm-base) - trained on complete s-PC-GITA and tested on enhanced e-PC-GITA
46
+ - [SSL4PR HuBERT Base Full](https://huggingface.co/morenolq/SSL4PR-hubert-base-full): using as base model [HuBERT Base](https://huggingface.co/facebook/hubert-base-ls960) - trained on complete s-PC-GITA and tested on enhanced e-PC-GITA
47
+
48
+ All models are available in PyTorch format.
49
+ ⚠️ Please note that the models are not directly compatible with the Hugging Face Transformers library because they are trained using specific head components (i.e., attention pooling, layer weighting...) as you can find in the [model class](https://github.com/K-STMLab/SSL4PR/blob/main/models/ssl_classification_model.py)
50
+
51
+ An image of the model architecture below:
52
 
53
+ ![Model Architecture](dspeech_arch.png)
 
54
 
55
  ## Citation
56