aifeifei798 commited on
Commit
8ee38b9
·
verified ·
1 Parent(s): d78f669

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +27 -15
app.py CHANGED
@@ -9,7 +9,10 @@ from diffusers import (
9
  AutoencoderTiny,
10
  )
11
  from huggingface_hub import hf_hub_download
12
-
 
 
 
13
 
14
  def feifeimodload():
15
 
@@ -36,11 +39,9 @@ def feifeimodload():
36
  return pipe
37
 
38
  pipe = feifeimodload()
39
- MAX_SEED = np.iinfo(np.int32).max
40
- MAX_IMAGE_SIZE = 2048
41
 
42
  @spaces.GPU()
43
- def infer(prompt="", Expansion_select=False, seed=42, randomize_seed=False, width=1024, height=1024, num_inference_steps=4, guidancescale=3.5, num_feifei=0.35, nsfw_select=False, nsfw_slider=1, progress=gr.Progress(track_tqdm=True)):
44
  if randomize_seed:
45
  seed = random.randint(0, MAX_SEED)
46
  generator = torch.Generator().manual_seed(seed)
@@ -73,7 +74,11 @@ def infer(prompt="", Expansion_select=False, seed=42, randomize_seed=False, widt
73
  prompt = re.sub(r"young woman", replacement, prompt, flags=re.IGNORECASE)
74
  prompt = re.sub(r"woman", replacement, prompt, flags=re.IGNORECASE)
75
  prompt = re.sub(r"model", replacement, prompt, flags=re.IGNORECASE)
76
-
 
 
 
 
77
  image = pipe(
78
  prompt = "flux, 8k, ",
79
  prompt_2 = prompt,
@@ -109,15 +114,22 @@ with gr.Blocks(css=css) as demo:
109
 
110
  with gr.Column(elem_id="col-container"):
111
  gr.Markdown(f"""# FLUX.1 + feifei-flux-lora """)
112
- prompt = gr.Text(
113
- label="Prompt",
114
- show_label=False,
115
- max_lines=12,
116
- placeholder="Enter your prompt",
117
- value="",
118
- container=False,
119
- )
120
-
 
 
 
 
 
 
 
121
  run_button = gr.Button("Run")
122
  result = gr.Image(label="Result", show_label=False,height=520)
123
  Expansion_select = gr.Checkbox(label="FeiFei Expansion")
@@ -200,7 +212,7 @@ with gr.Blocks(css=css) as demo:
200
  gr.on(
201
  triggers=[run_button.click, prompt.submit],
202
  fn = infer,
203
- inputs = [prompt, Expansion_select, seed, randomize_seed, width, height, num_inference_steps, guidancescale, num_feifei, nsfw_select, nsfw_slider],
204
  outputs = [result, seed]
205
  )
206
 
 
9
  AutoencoderTiny,
10
  )
11
  from huggingface_hub import hf_hub_download
12
+ import config
13
+ styles_name = [style["name"] for style in config.style_list]
14
+ MAX_SEED = np.iinfo(np.int32).max
15
+ MAX_IMAGE_SIZE = 2048
16
 
17
  def feifeimodload():
18
 
 
39
  return pipe
40
 
41
  pipe = feifeimodload()
 
 
42
 
43
  @spaces.GPU()
44
+ def infer(prompt="", styles_Radio="(None)", Expansion_select=False, seed=42, randomize_seed=False, width=1024, height=1024, num_inference_steps=4, guidancescale=3.5, num_feifei=0.35, nsfw_select=False, nsfw_slider=1, progress=gr.Progress(track_tqdm=True)):
45
  if randomize_seed:
46
  seed = random.randint(0, MAX_SEED)
47
  generator = torch.Generator().manual_seed(seed)
 
74
  prompt = re.sub(r"young woman", replacement, prompt, flags=re.IGNORECASE)
75
  prompt = re.sub(r"woman", replacement, prompt, flags=re.IGNORECASE)
76
  prompt = re.sub(r"model", replacement, prompt, flags=re.IGNORECASE)
77
+ if styles_Radio:
78
+ style_name = styles_Radio
79
+ for style in config.style_list:
80
+ if style["name"] == style_name:
81
+ prompt = style["prompt"].replace("{prompt}", prompt)
82
  image = pipe(
83
  prompt = "flux, 8k, ",
84
  prompt_2 = prompt,
 
114
 
115
  with gr.Column(elem_id="col-container"):
116
  gr.Markdown(f"""# FLUX.1 + feifei-flux-lora """)
117
+ with gr.Row():
118
+ prompt = gr.Text(
119
+ label="Prompt",
120
+ show_label=False,
121
+ max_lines=12,
122
+ placeholder="Enter your prompt",
123
+ value="",
124
+ container=False,
125
+ )
126
+ with gr.Row():
127
+ styles_Radio = gr.Dropdown(
128
+ styles_name,
129
+ label="Styles",
130
+ multiselect=False,
131
+ value="(None)",
132
+ )
133
  run_button = gr.Button("Run")
134
  result = gr.Image(label="Result", show_label=False,height=520)
135
  Expansion_select = gr.Checkbox(label="FeiFei Expansion")
 
212
  gr.on(
213
  triggers=[run_button.click, prompt.submit],
214
  fn = infer,
215
+ inputs = [prompt, styles_Radio, Expansion_select, seed, randomize_seed, width, height, num_inference_steps, guidancescale, num_feifei, nsfw_select, nsfw_slider],
216
  outputs = [result, seed]
217
  )
218