LAP-DEV commited on
Commit
007979f
·
verified ·
1 Parent(s): a712f8a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +40 -37
app.py CHANGED
@@ -124,43 +124,46 @@ class App:
124
 
125
  with gr.Row():
126
  with gr.Column():
127
- with gr.Accordion("Voice Detection Filter (⚠ experimental feature)", open=False, visible=True):
128
- cb_vad_filter = gr.Checkbox(label="Enable Silero VAD Filter", value=vad_params["vad_filter"],
129
- interactive=True,
130
- info="Enable to transcribe only detected voice parts")
131
- sd_threshold = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label="Speech Threshold",
132
- value=vad_params["threshold"],
133
- info="Lower it to be more sensitive to small sounds")
134
- nb_min_speech_duration_ms = gr.Number(label="Minimum Speech Duration (ms)", precision=0,
135
- value=vad_params["min_speech_duration_ms"],
136
- info="Final speech chunks shorter than this time are thrown out")
137
- nb_max_speech_duration_s = gr.Number(label="Maximum Speech Duration (s)",
138
- value=vad_params["max_speech_duration_s"],
139
- info="Maximum duration of speech chunks in seconds")
140
- nb_min_silence_duration_ms = gr.Number(label="Minimum Silence Duration (ms)", precision=0,
141
- value=vad_params["min_silence_duration_ms"],
142
- info="In the end of each speech chunk wait for this time"
143
- " before separating it")
144
- nb_speech_pad_ms = gr.Number(label="Speech Padding (ms)", precision=0, value=vad_params["speech_pad_ms"],
145
- info="Final speech chunks are padded by this time each side")
146
-
147
-
148
- with gr.Accordion("Background Music Remover Filter (⚠ experimental feature)", open=False):
149
- cb_bgm_separation = gr.Checkbox(label="Enable Background Music Remover Filter", value=uvr_params["is_separate_bgm"],
150
- info="Enable to remove background music by submodel before transcribing",
151
- interactive=True)
152
- dd_uvr_device = gr.Dropdown(label="Device",
153
- value=self.whisper_inf.music_separator.device,
154
- choices=self.whisper_inf.music_separator.available_devices,
155
- interactive=True, visible=False)
156
- dd_uvr_model_size = gr.Dropdown(label="Model", value=uvr_params["model_size"],
157
- choices=self.whisper_inf.music_separator.available_models)
158
- nb_uvr_segment_size = gr.Number(label="Segment Size", value=uvr_params["segment_size"], precision=0,
159
- interactive=True, visible=False)
160
- cb_uvr_save_file = gr.Checkbox(label="Save separated files to output", value=uvr_params["save_file"],
161
- interactive=True, visible=False)
162
- cb_uvr_enable_offload = gr.Checkbox(label="Offload sub model after removing background music",value=uvr_params["enable_offload"],
163
- interactive=True, visible=False)
 
 
 
164
 
165
  with gr.Accordion("Advanced processing options", open=False, visible=False):
166
  nb_beam_size = gr.Number(label="Beam Size", value=whisper_params["beam_size"], precision=0, interactive=True,
 
124
 
125
  with gr.Row():
126
  with gr.Column():
127
+ with gr.Row():
128
+ gr.Markdown(" experimental features sdf dsfdsf sdfsd sdfsd")
129
+ with gr.Row():
130
+ with gr.Accordion("Voice Detection Filter (⚠ experimental feature)", open=False, visible=True):
131
+ cb_vad_filter = gr.Checkbox(label="Enable Silero VAD Filter", value=vad_params["vad_filter"],
132
+ interactive=True,
133
+ info="Enable to transcribe only detected voice parts")
134
+ sd_threshold = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label="Speech Threshold",
135
+ value=vad_params["threshold"],
136
+ info="Lower it to be more sensitive to small sounds")
137
+ nb_min_speech_duration_ms = gr.Number(label="Minimum Speech Duration (ms)", precision=0,
138
+ value=vad_params["min_speech_duration_ms"],
139
+ info="Final speech chunks shorter than this time are thrown out")
140
+ nb_max_speech_duration_s = gr.Number(label="Maximum Speech Duration (s)",
141
+ value=vad_params["max_speech_duration_s"],
142
+ info="Maximum duration of speech chunks in seconds")
143
+ nb_min_silence_duration_ms = gr.Number(label="Minimum Silence Duration (ms)", precision=0,
144
+ value=vad_params["min_silence_duration_ms"],
145
+ info="In the end of each speech chunk wait for this time"
146
+ " before separating it")
147
+ nb_speech_pad_ms = gr.Number(label="Speech Padding (ms)", precision=0, value=vad_params["speech_pad_ms"],
148
+ info="Final speech chunks are padded by this time each side")
149
+
150
+
151
+ with gr.Accordion("Background Music Remover Filter (⚠ experimental feature)", open=False):
152
+ cb_bgm_separation = gr.Checkbox(label="Enable Background Music Remover Filter", value=uvr_params["is_separate_bgm"],
153
+ info="Enable to remove background music by submodel before transcribing",
154
+ interactive=True)
155
+ dd_uvr_device = gr.Dropdown(label="Device",
156
+ value=self.whisper_inf.music_separator.device,
157
+ choices=self.whisper_inf.music_separator.available_devices,
158
+ interactive=True, visible=False)
159
+ dd_uvr_model_size = gr.Dropdown(label="Model", value=uvr_params["model_size"],
160
+ choices=self.whisper_inf.music_separator.available_models)
161
+ nb_uvr_segment_size = gr.Number(label="Segment Size", value=uvr_params["segment_size"], precision=0,
162
+ interactive=True, visible=False)
163
+ cb_uvr_save_file = gr.Checkbox(label="Save separated files to output", value=uvr_params["save_file"],
164
+ interactive=True, visible=False)
165
+ cb_uvr_enable_offload = gr.Checkbox(label="Offload sub model after removing background music",value=uvr_params["enable_offload"],
166
+ interactive=True, visible=False)
167
 
168
  with gr.Accordion("Advanced processing options", open=False, visible=False):
169
  nb_beam_size = gr.Number(label="Beam Size", value=whisper_params["beam_size"], precision=0, interactive=True,