Marianoleiras commited on
Commit
23926f1
·
verified ·
1 Parent(s): 7916cef

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +12 -14
README.md CHANGED
@@ -7,6 +7,13 @@ metrics:
7
  model-index:
8
  - name: whisper-small-es-ja
9
  results: []
 
 
 
 
 
 
 
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -14,7 +21,10 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # whisper-small-es-ja
16
 
17
- This model was trained from scratch on an unknown dataset.
 
 
 
18
  It achieves the following results on the evaluation set:
19
  - Loss: 1.1724
20
  - Bleu: 22.2850
@@ -22,18 +32,6 @@ It achieves the following results on the evaluation set:
22
  It achieves the following results on the test set:
23
  - Bleu: 21.4557
24
 
25
- ## Model description
26
-
27
- More information needed
28
-
29
- ## Intended uses & limitations
30
-
31
- More information needed
32
-
33
- ## Training and evaluation data
34
-
35
- More information needed
36
-
37
  ## Training procedure
38
 
39
  ### Training hyperparameters
@@ -74,4 +72,4 @@ The following hyperparameters were used during training:
74
  - Transformers 4.47.1
75
  - Pytorch 2.4.0+cu124
76
  - Datasets 3.2.0
77
- - Tokenizers 0.21.0
 
7
  model-index:
8
  - name: whisper-small-es-ja
9
  results: []
10
+ datasets:
11
+ - Marianoleiras/voxpopuli_es-ja
12
+ language:
13
+ - es
14
+ - ja
15
+ base_model:
16
+ - openai/whisper-small
17
  ---
18
 
19
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
21
 
22
  # whisper-small-es-ja
23
 
24
+ This model is a fine-tuned version of OpenAI's whisper-small on the Marianoleiras/voxpopuli_es-ja dataset, designed for Spanish-to-Japanese and Japanese-to-Spanish speech-to-text (STT) tasks.
25
+ It leverages OpenAI's Whisper architecture, which is well-suited for multilingual speech recognition and translation tasks.
26
+ The model achieves robust performance on both the evaluation and test sets, demonstrating its effectiveness in multilingual STT applications.
27
+
28
  It achieves the following results on the evaluation set:
29
  - Loss: 1.1724
30
  - Bleu: 22.2850
 
32
  It achieves the following results on the test set:
33
  - Bleu: 21.4557
34
 
 
 
 
 
 
 
 
 
 
 
 
 
35
  ## Training procedure
36
 
37
  ### Training hyperparameters
 
72
  - Transformers 4.47.1
73
  - Pytorch 2.4.0+cu124
74
  - Datasets 3.2.0
75
+ - Tokenizers 0.21.0