Update app.py
Browse files
app.py
CHANGED
@@ -101,13 +101,15 @@ class App:
|
|
101 |
input_multi = gr.Radio(["Audio", "Video", "Multiple"], label="Process one or multiple files", value="Audio")
|
102 |
with gr.Row():
|
103 |
dd_file_format = gr.Dropdown(choices=["CSV","SRT","TXT"], value=whisper_params["output_format"], label="Output format", multiselect=True, interactive=True, visible=True)
|
|
|
|
|
104 |
with gr.Column(scale=4):
|
105 |
input_file_audio = gr.Audio(type='filepath', elem_id="audio_input", show_download_button=True, visible=True, interactive=True)
|
106 |
input_file_video = gr.Video(elem_id="video_input", show_download_button=True, visible=False, interactive=True)
|
107 |
input_file_multi = gr.Files(label="Upload one or more audio/video files here", elem_id="multi_input", type='filepath', file_count="multiple", allow_reordering=True, file_types=["audio","video"], visible=False, interactive=True)
|
108 |
|
109 |
with gr.Row():
|
110 |
-
with gr.Column(scale=
|
111 |
with gr.Row():
|
112 |
model_list = self.whisper_inf.available_models if len(whisper_params["model_list"]) == 0 else whisper_params["model_list"]
|
113 |
dd_model = gr.Dropdown(choices=model_list, value=whisper_params["model_size"],label="Model", info="Larger models increase transcription quality, but reduce performance", interactive=True)
|
@@ -116,12 +118,8 @@ class App:
|
|
116 |
dd_translate_model = gr.Dropdown(choices=self.nllb_inf.available_models, value=nllb_params["model_size"],label="Model", info="Model used for translation", interactive=True)
|
117 |
dd_target_lang = gr.Dropdown(choices=["English","Dutch","French","German"], value=nllb_params["target_lang"],label="Language", info="Language used for output translation", interactive=True)
|
118 |
with gr.Column(scale=1):
|
119 |
-
with gr.Row():
|
120 |
-
cb_timestamp_preview = gr.Checkbox(value=whisper_params["add_timestamp_preview"],label="Add timestamp to preview", interactive=True)
|
121 |
-
cb_timestamp_file = gr.Checkbox(value=whisper_params["add_timestamp_file"], label="Add timestamp to output", interactive=True)
|
122 |
with gr.Row():
|
123 |
cb_diarize = gr.Checkbox(value=diarization_params["is_diarize"],label="Speaker diarization", info="Token required",interactive=True)
|
124 |
-
with gr.Column(scale=1):
|
125 |
with gr.Row():
|
126 |
cb_translate = gr.Checkbox(value=whisper_params["is_translate"], label="Translate to English", info="Translate using OpenAI Whisper's built-in module",interactive=True)
|
127 |
cb_translate_output = gr.Checkbox(value=translation_params["translate_output"], label="Translate to selected language", info="Translate using Facebook's NLLB",interactive=True)
|
|
|
101 |
input_multi = gr.Radio(["Audio", "Video", "Multiple"], label="Process one or multiple files", value="Audio")
|
102 |
with gr.Row():
|
103 |
dd_file_format = gr.Dropdown(choices=["CSV","SRT","TXT"], value=whisper_params["output_format"], label="Output format", multiselect=True, interactive=True, visible=True)
|
104 |
+
cb_timestamp_preview = gr.Checkbox(value=whisper_params["add_timestamp_preview"],label="Add timestamp to preview", interactive=True)
|
105 |
+
cb_timestamp_file = gr.Checkbox(value=whisper_params["add_timestamp_file"], label="Add timestamp to output", interactive=True)
|
106 |
with gr.Column(scale=4):
|
107 |
input_file_audio = gr.Audio(type='filepath', elem_id="audio_input", show_download_button=True, visible=True, interactive=True)
|
108 |
input_file_video = gr.Video(elem_id="video_input", show_download_button=True, visible=False, interactive=True)
|
109 |
input_file_multi = gr.Files(label="Upload one or more audio/video files here", elem_id="multi_input", type='filepath', file_count="multiple", allow_reordering=True, file_types=["audio","video"], visible=False, interactive=True)
|
110 |
|
111 |
with gr.Row():
|
112 |
+
with gr.Column(scale=4):
|
113 |
with gr.Row():
|
114 |
model_list = self.whisper_inf.available_models if len(whisper_params["model_list"]) == 0 else whisper_params["model_list"]
|
115 |
dd_model = gr.Dropdown(choices=model_list, value=whisper_params["model_size"],label="Model", info="Larger models increase transcription quality, but reduce performance", interactive=True)
|
|
|
118 |
dd_translate_model = gr.Dropdown(choices=self.nllb_inf.available_models, value=nllb_params["model_size"],label="Model", info="Model used for translation", interactive=True)
|
119 |
dd_target_lang = gr.Dropdown(choices=["English","Dutch","French","German"], value=nllb_params["target_lang"],label="Language", info="Language used for output translation", interactive=True)
|
120 |
with gr.Column(scale=1):
|
|
|
|
|
|
|
121 |
with gr.Row():
|
122 |
cb_diarize = gr.Checkbox(value=diarization_params["is_diarize"],label="Speaker diarization", info="Token required",interactive=True)
|
|
|
123 |
with gr.Row():
|
124 |
cb_translate = gr.Checkbox(value=whisper_params["is_translate"], label="Translate to English", info="Translate using OpenAI Whisper's built-in module",interactive=True)
|
125 |
cb_translate_output = gr.Checkbox(value=translation_params["translate_output"], label="Translate to selected language", info="Translate using Facebook's NLLB",interactive=True)
|