Update app.py
Browse files
app.py
CHANGED
@@ -30,13 +30,13 @@ with open(diffusion_config, 'r') as fp:
|
|
30 |
|
31 |
v_prediction = diff_config["ddim"]["v_prediction"]
|
32 |
|
33 |
-
# clapmodel = ClapModel.from_pretrained("laion/larger_clap_general").to(device)
|
34 |
processor = AutoProcessor.from_pretrained('laion/larger_clap_general')
|
35 |
clap_config = ClapConfig.from_pretrained("laion/larger_clap_general")
|
36 |
clapmodel = ClapModel(config)
|
37 |
clap_ckpt = torch.load(clap_bin_path, map_location='cpu')
|
38 |
clapmodel.load_state_dict(clap_ckpt)
|
39 |
clapmodel.to(device)
|
|
|
40 |
|
41 |
autoencoder = Autoencoder(autoencoder_path, 'stable_vae', quantization_first=True)
|
42 |
autoencoder.eval()
|
|
|
30 |
|
31 |
v_prediction = diff_config["ddim"]["v_prediction"]
|
32 |
|
|
|
33 |
processor = AutoProcessor.from_pretrained('laion/larger_clap_general')
|
34 |
clap_config = ClapConfig.from_pretrained("laion/larger_clap_general")
|
35 |
clapmodel = ClapModel(config)
|
36 |
clap_ckpt = torch.load(clap_bin_path, map_location='cpu')
|
37 |
clapmodel.load_state_dict(clap_ckpt)
|
38 |
clapmodel.to(device)
|
39 |
+
# clapmodel = ClapModel.from_pretrained("laion/larger_clap_general").to(device)
|
40 |
|
41 |
autoencoder = Autoencoder(autoencoder_path, 'stable_vae', quantization_first=True)
|
42 |
autoencoder.eval()
|