File size: 2,400 Bytes
c19ca42
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
import os
import argparse
import torch
from diffusers import StableDiffusionPipeline, StableDiffusionXLPipeline, AutoPipelineForText2Image, LCMScheduler

parser = argparse.ArgumentParser("lcm_convert")
parser.add_argument("--name", help="Name of the new LCM model", type=str)
parser.add_argument("--model", help="A model to convert", type=str)
parser.add_argument("--lora-scale", default=1.0, help="Strenght of the LCM", type=float)
parser.add_argument("--huggingface", action="store_true", help="Use Hugging Face models instead of safetensors models")
parser.add_argument("--upload", action="store_true", help="Upload the new LCM model to Hugging Face")
parser.add_argument("--no-half", action="store_true", help="Convert the new LCM model to FP32")
parser.add_argument("--no-save", action="store_true", help="Don't save the new LCM model to local disk")
parser.add_argument("--sdxl", action="store_true", help="Use SDXL models")
parser.add_argument("--ssd-1b", action="store_true", help="Use SSD-1B models")

args = parser.parse_args()

if args.huggingface:
    pipeline = AutoPipelineForText2Image.from_pretrained(args.model, torch_dtype=torch.float16, variant="fp16")
else:
    if args.sdxl or args.ssd_1b:
        pipeline = StableDiffusionXLPipeline.from_single_file(args.model)
    else:
        pipeline = StableDiffusionPipeline.from_single_file(args.model)

pipeline.scheduler = LCMScheduler.from_config(pipeline.scheduler.config)
if args.sdxl:
    pipeline.load_lora_weights("latent-consistency/lcm-lora-sdxl")
elif args.ssd_1b:
    pipeline.load_lora_weights("latent-consistency/lcm-lora-ssd-1b")
else:
    pipeline.load_lora_weights("latent-consistency/lcm-lora-sdv1-5")
pipeline.fuse_lora(lora_scale=args.lora_scale)

#components = pipeline.components
#pipeline = LatentConsistencyModelPipeline(**components)

if args.no_half:
    pipeline = pipeline.to(dtype=torch.float32)
else:
    pipeline = pipeline.to(dtype=torch.float16)
print(pipeline)

if not args.no_save:
    os.makedirs(f"models--local--{args.name}/snapshots")
    if args.no_half:
        pipeline.save_pretrained(f"models--local--{args.name}/snapshots/{args.name}")
    else:
        pipeline.save_pretrained(f"models--local--{args.name}/snapshots/{args.name}", variant="fp16")
if args.upload:
    if args.no_half:
        pipeline.push_to_hub(args.name)
    else:
        pipeline.push_to_hub(args.name, variant="fp16")