Spaces:
Sleeping
Sleeping
import gradio as gr | |
import numpy as np | |
import torch | |
from PIL import Image | |
from diffusers import StableDiffusionPipeline | |
from transformers import pipeline, set_seed | |
import random | |
import re | |
model_id = "stable-diffusion-v1-5/stable-diffusion-v1-5" | |
pipe = StableDiffusionPipeline.from_pretrained(model_id).to('cpu') | |
gpt2_pipe = pipeline('text-generation', model='Gustavosta/MagicPrompt-Stable-Diffusion', tokenizer='gpt2') | |
gpt2_pipe2 = pipeline('text-generation', model='succinctly/text2image-prompt-generator') | |
def infer1(starting_text): | |
seed = random.randint(100, 1000000) | |
set_seed(seed) | |
if starting_text == "": | |
starting_text: str = re.sub(r"[,:\-β.!;?_]", '', starting_text) | |
response = gpt2_pipe(starting_text, max_length=(len(starting_text) + random.randint(60, 90)), num_return_sequences=4) | |
response_list = [] | |
for x in response: | |
resp = x['generated_text'].strip() | |
if resp != starting_text and len(resp) > (len(starting_text) + 4) and resp.endswith((":", "-", "β")) is False: | |
response_list.append(resp+'\n') | |
response_end = "\n".join(response_list) | |
response_end = re.sub('[^ ]+\.[^ ]+','', response_end) | |
response_end = response_end.replace("<", "").replace(">", "") | |
if response_end != "": | |
return response_end | |
def infer2(starting_text): | |
for count in range(6): | |
seed = random.randint(100, 1000000) | |
set_seed(seed) | |
# If the text field is empty | |
if starting_text == "": | |
starting_text: str = line[random.randrange(0, len(line))].replace("\n", "").lower().capitalize() | |
starting_text: str = re.sub(r"[,:\-β.!;?_]", '', starting_text) | |
print(starting_text) | |
response = gpt2_pipe2(starting_text, max_length=random.randint(60, 90), num_return_sequences=8) | |
response_list = [] | |
for x in response: | |
resp = x['generated_text'].strip() | |
if resp != starting_text and len(resp) > (len(starting_text) + 4) and resp.endswith((":", "-", "β")) is False: | |
response_list.append(resp) | |
response_end = "\n".join(response_list) | |
response_end = re.sub('[^ ]+\.[^ ]+','', response_end) | |
response_end = response_end.replace("<", "").replace(">", "") | |
if response_end != "": | |
return response_end | |
if count == 5: | |
return response_end | |
def infer3(prompt, negative, steps, scale, seed): | |
generator = torch.Generator(device='cpu').manual_seed(seed) | |
img = pipe( | |
prompt, | |
height=512, | |
width=512, | |
num_inference_steps=steps, | |
guidance_scale=scale, | |
negative_prompt = negative, | |
generator=generator, | |
).images | |
return img | |
block = gr.Blocks() | |
with block: | |
with gr.Group(): | |
with gr.Box(): | |
gr.Markdown( | |
""" | |
Model: Gustavosta/MagicPrompt-Stable-Diffusion | |
""" | |
) | |
with gr.Row() as row: | |
with gr.Column(): | |
txt = gr.Textbox(lines=1, label="Initial Text", placeholder="English Text here") | |
gpt_btn = gr.Button("Generate prompt").style( | |
margin=False, | |
rounded=(False, True, True, False), | |
) | |
with gr.Column(): | |
out = gr.Textbox(lines=4, label="Generated Prompts") | |
with gr.Box(): | |
gr.Markdown( | |
""" | |
Model: succinctly/text2image-prompt-generator | |
""" | |
) | |
with gr.Row() as row: | |
with gr.Column(): | |
txt2 = gr.Textbox(lines=1, label="Initial Text", placeholder="English Text here") | |
gpt_btn2 = gr.Button("Generate prompt").style( | |
margin=False, | |
rounded=(False, True, True, False), | |
) | |
with gr.Column(): | |
out2 = gr.Textbox(lines=4, label="Generated Prompts") | |
with gr.Box(): | |
gr.Markdown( | |
""" | |
Model: stable diffusion v1.5 | |
""" | |
) | |
with gr.Row(elem_id="prompt-container").style(mobile_collapse=False, equal_height=True): | |
with gr.Column(): | |
text = gr.Textbox( | |
label="Enter your prompt", | |
show_label=False, | |
max_lines=1, | |
placeholder="Enter your prompt", | |
).style( | |
border=(True, False, True, True), | |
rounded=(True, False, False, True), | |
container=False, | |
) | |
negative = gr.Textbox( | |
label="Enter your negative prompt", | |
show_label=False, | |
placeholder="Enter a negative prompt", | |
elem_id="negative-prompt-text-input", | |
).style( | |
border=(True, False, True, True), | |
rounded=(True, False, False, True),container=False, | |
) | |
btn = gr.Button("Generate image").style( | |
margin=False, | |
rounded=(False, True, True, False), | |
) | |
gallery = gr.Gallery( | |
label="Generated images", show_label=False, elem_id="gallery" | |
).style(columns=(1, 2), height="auto") | |
with gr.Row(elem_id="advanced-options"): | |
samples = gr.Slider(label="Images", minimum=1, maximum=1, value=1, step=1, interactive=False) | |
steps = gr.Slider(label="Steps", minimum=1, maximum=50, value=12, step=1, interactive=True) | |
scale = gr.Slider(label="Guidance Scale", minimum=0, maximum=50, value=7.5, step=0.1, interactive=True) | |
seed = gr.Slider(label="Random seed",minimum=0,maximum=2147483647,step=1,randomize=True,interactive=True) | |
gpt_btn.click(infer1,inputs=txt,outputs=out) | |
gpt_btn2.click(infer2,inputs=txt2,outputs=out2) | |
btn.click(infer3, inputs=[text, negative, steps, scale, seed], outputs=[gallery]) | |
block.launch(show_api=False,enable_queue=True, debug=True) |