Update README.md
Browse files
README.md
CHANGED
@@ -10,6 +10,68 @@ language:
|
|
10 |
- en
|
11 |
---
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
|
14 |
-
- Library: [
|
15 |
-
-
|
|
|
10 |
- en
|
11 |
---
|
12 |
|
13 |
+
### 1. Introduction
|
14 |
+
Transformers and their variants have achieved great
|
15 |
+
success in speech processing. However, their multi-head selfattention mechanism is computationally expensive. Therefore, one
|
16 |
+
novel selective state space model, Mamba, has been proposed
|
17 |
+
as an alternative. Building on its success in automatic speech
|
18 |
+
recognition, we apply Mamba for spoofing attack detection.
|
19 |
+
Mamba is well-suited for this task as it can capture the artifacts
|
20 |
+
in spoofed speech signals by handling long-length sequences.
|
21 |
+
However, Mamba’s performance may suffer when it is trained
|
22 |
+
with limited labeled data. To mitigate this, we propose combining
|
23 |
+
a new structure of Mamba based on a dual-column architecture
|
24 |
+
with self-supervised learning, using the pre-trained wav2vec
|
25 |
+
2.0 model. The experiments show that our proposed approach
|
26 |
+
achieves competitive results and faster inference on the ASVspoof
|
27 |
+
2021 LA and DF datasets, and on the more challenging In-theWild dataset, it emerges as the strongest candidate for spoofing
|
28 |
+
attack detection.
|
29 |
+
|
30 |
+
### 2. Setup Environment
|
31 |
+
You need to create the running environment by [Anaconda](https://www.anaconda.com/).
|
32 |
+
First, create and activate the environment:
|
33 |
+
|
34 |
+
```bash
|
35 |
+
conda create -n XLSR_Mamba python=3.10
|
36 |
+
conda activate XLSR_Mamba
|
37 |
+
```
|
38 |
+
|
39 |
+
Then install the requirements:
|
40 |
+
|
41 |
+
```bash
|
42 |
+
pip install -r requirements.txt
|
43 |
+
```
|
44 |
+
|
45 |
+
Install fairseq:
|
46 |
+
|
47 |
+
```bash
|
48 |
+
git clone https://github.com/facebookresearch/fairseq.git fairseq_dir
|
49 |
+
cd fairseq_dir
|
50 |
+
git checkout a54021305d6b3c
|
51 |
+
pip install --editable ./
|
52 |
+
```
|
53 |
+
### 3. Pretrained Model
|
54 |
+
The pretrained model XLSR can be found at this [link](https://dl.fbaipublicfiles.com/fairseq/wav2vec/xlsr2_300m.pt).
|
55 |
+
|
56 |
+
### 4. Results
|
57 |
+
|
58 |
+
|
59 |
+
Dataset | **EER (%)** | **min t-DCF** |
|
60 |
+
--------|--------------------|--------------------|
|
61 |
+
ASVspoof2021 LA | **0.93** | **0.208** |
|
62 |
+
ASVspoof2021 DF | **1.88** | **-** |
|
63 |
+
In-The-Wild | **6.71** | **-** |
|
64 |
+
|
65 |
+
### 5. Citation
|
66 |
+
If you find our repository valuable for your work, please consider giving a star to this repo and citing our paper:
|
67 |
+
```
|
68 |
+
@article{xiao2024xlsr,
|
69 |
+
title={{XLSR-Mamba}: A Dual-Column Bidirectional State Space Model for Spoofing Attack Detection},
|
70 |
+
author={Xiao, Yang and Das, Rohan Kumar},
|
71 |
+
journal={arXiv preprint arXiv:2411.10027},
|
72 |
+
year={2024}
|
73 |
+
}
|
74 |
+
```
|
75 |
This model has been pushed to the Hub using the [PytorchModelHubMixin](https://huggingface.co/docs/huggingface_hub/package_reference/mixins#huggingface_hub.PyTorchModelHubMixin) integration:
|
76 |
+
- Library: [https://github.com/swagshaw/XLSR-Mamba]
|
77 |
+
- Paper: [https://arxiv.org/pdf/2411.10027]
|