Jenthe commited on
Commit
e49fc6e
·
verified ·
1 Parent(s): e7f1142

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -82,6 +82,7 @@ with torch.jit.optimized_execution(False):
82
 
83
  There is no need for `ecapa2.eval()` or `torch.no_grad()`, this is done automatically.
84
 
 
85
  ### Hierarchical Feature Extraction
86
 
87
  For the extraction of other hierachical features, the `label` argument can be used, which accepts a string containing the feature ids separated with '|':
@@ -106,7 +107,7 @@ The following table describes the available features. All features consists of t
106
  | pool | 3072 | Pooled statistics before the bottleneck speaker embedding layer, extracted before ReLU layer.
107
  | attention | 3072 | Same as the pooled statistics but with the attention weights applied.
108
  | embedding | 192 | The standard ECAPA2 speaker embedding.
109
- <!--
110
  The following table describes the available features:
111
 
112
  | Feature Type| Description | Usage | Labels |
@@ -114,8 +115,8 @@ The following table describes the available features:
114
  | Local Feature | Non-uniform effective receptive field in the frequency dimension of each frame-level feature.| Abstract features, probably usefull in tasks less related to speaker characteristics. | lfe1, lfe2, lfe3, lfe4
115
  | Global Feature | Uniform effective receptive field of each frame-level feature in the frequency dimension.| Generally capture intra-speaker variance better then speaker embeddings. E.g. speaker profiling, emotion recognition. | gfe1, gfe2, gfe3, pool
116
  | Speaker Embedding | Uniform effective receptive field of each frame-level feature in the frequency dimension.| Best for tasks directly depending on the speaker identity (as opposed to speaker characteristics). E.g. speaker verification, speaker diarization. | embedding
 
117
  -->
118
- <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
119
 
120
  ## Citation
121
 
 
82
 
83
  There is no need for `ecapa2.eval()` or `torch.no_grad()`, this is done automatically.
84
 
85
+ <!--
86
  ### Hierarchical Feature Extraction
87
 
88
  For the extraction of other hierachical features, the `label` argument can be used, which accepts a string containing the feature ids separated with '|':
 
107
  | pool | 3072 | Pooled statistics before the bottleneck speaker embedding layer, extracted before ReLU layer.
108
  | attention | 3072 | Same as the pooled statistics but with the attention weights applied.
109
  | embedding | 192 | The standard ECAPA2 speaker embedding.
110
+
111
  The following table describes the available features:
112
 
113
  | Feature Type| Description | Usage | Labels |
 
115
  | Local Feature | Non-uniform effective receptive field in the frequency dimension of each frame-level feature.| Abstract features, probably usefull in tasks less related to speaker characteristics. | lfe1, lfe2, lfe3, lfe4
116
  | Global Feature | Uniform effective receptive field of each frame-level feature in the frequency dimension.| Generally capture intra-speaker variance better then speaker embeddings. E.g. speaker profiling, emotion recognition. | gfe1, gfe2, gfe3, pool
117
  | Speaker Embedding | Uniform effective receptive field of each frame-level feature in the frequency dimension.| Best for tasks directly depending on the speaker identity (as opposed to speaker characteristics). E.g. speaker verification, speaker diarization. | embedding
118
+
119
  -->
 
120
 
121
  ## Citation
122