Update app_v1v2.py
Browse files- app_v1v2.py +4 -4
app_v1v2.py
CHANGED
@@ -23,10 +23,6 @@ if hasattr(torch._inductor.config, "fx_graph_cache"):
|
|
23 |
|
24 |
dtype = torch.float32
|
25 |
|
26 |
-
# Global variables to store model instances
|
27 |
-
vc_wrapper_v1 = SeedVCWrapper()
|
28 |
-
vc_wrapper_v2 = load_v2_models(args)
|
29 |
-
|
30 |
|
31 |
def load_v2_models():
|
32 |
from hydra.utils import instantiate
|
@@ -41,6 +37,10 @@ def load_v2_models():
|
|
41 |
|
42 |
return vc_wrapper
|
43 |
|
|
|
|
|
|
|
|
|
44 |
@spaces.GPU
|
45 |
def convert_voice_v1_wrapper(source_audio_path, target_audio_path, diffusion_steps=10,
|
46 |
length_adjust=1.0, inference_cfg_rate=0.7, f0_condition=False,
|
|
|
23 |
|
24 |
dtype = torch.float32
|
25 |
|
|
|
|
|
|
|
|
|
26 |
|
27 |
def load_v2_models():
|
28 |
from hydra.utils import instantiate
|
|
|
37 |
|
38 |
return vc_wrapper
|
39 |
|
40 |
+
# Global variables to store model instances
|
41 |
+
vc_wrapper_v1 = SeedVCWrapper()
|
42 |
+
vc_wrapper_v2 = load_v2_models(args)
|
43 |
+
|
44 |
@spaces.GPU
|
45 |
def convert_voice_v1_wrapper(source_audio_path, target_audio_path, diffusion_steps=10,
|
46 |
length_adjust=1.0, inference_cfg_rate=0.7, f0_condition=False,
|