multimodalart HF Staff commited on
Commit
f6b6dd7
Β·
verified Β·
1 Parent(s): a8ca9e4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -330,6 +330,9 @@ css = """
330
  margin: 0 auto;
331
  max-width: 800px;
332
  }
 
 
 
333
  """
334
  example = """
335
  Emotion Description
@@ -377,7 +380,7 @@ Following the algorithm proposed in IP-Composer: Semantic Composition of Visual
377
  with gr.Tab("Concept 1"):
378
  with gr.Group():
379
  concept_image1 = gr.Image(label="Concept Image 1", type="numpy", height=400, width=400)
380
- with gr.Row():
381
  concept_name1 = gr.Dropdown(concept_options, label="Concept 1", value=None, info="Pick concept type")
382
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
383
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")
@@ -389,7 +392,7 @@ Following the algorithm proposed in IP-Composer: Semantic Composition of Visual
389
  with gr.Tab("Concept 2 (Optional)"):
390
  with gr.Group():
391
  concept_image2 = gr.Image(label="Concept Image 2", type="numpy", height=400, width=400)
392
- with gr.Row():
393
  concept_name2 = gr.Dropdown(concept_options, label="Concept 2", value=None, info="Pick concept type")
394
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
395
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")
@@ -402,7 +405,7 @@ Following the algorithm proposed in IP-Composer: Semantic Composition of Visual
402
  with gr.Tab("Concept 3 (optional)"):
403
  with gr.Group():
404
  concept_image3 = gr.Image(label="Concept Image 3", type="numpy", height=400, width=400)
405
- with gr.Row():
406
  concept_name3 = gr.Dropdown(concept_options, label="Concept 3", value= None, info="Pick concept type")
407
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
408
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")
 
330
  margin: 0 auto;
331
  max-width: 800px;
332
  }
333
+ .gradio-container{
334
+ max-width: 1024px !important
335
+ }
336
  """
337
  example = """
338
  Emotion Description
 
380
  with gr.Tab("Concept 1"):
381
  with gr.Group():
382
  concept_image1 = gr.Image(label="Concept Image 1", type="numpy", height=400, width=400)
383
+ with gr.Column():
384
  concept_name1 = gr.Dropdown(concept_options, label="Concept 1", value=None, info="Pick concept type")
385
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
386
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")
 
392
  with gr.Tab("Concept 2 (Optional)"):
393
  with gr.Group():
394
  concept_image2 = gr.Image(label="Concept Image 2", type="numpy", height=400, width=400)
395
+ with gr.Column():
396
  concept_name2 = gr.Dropdown(concept_options, label="Concept 2", value=None, info="Pick concept type")
397
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
398
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")
 
405
  with gr.Tab("Concept 3 (optional)"):
406
  with gr.Group():
407
  concept_image3 = gr.Image(label="Concept Image 3", type="numpy", height=400, width=400)
408
+ with gr.Column():
409
  concept_name3 = gr.Dropdown(concept_options, label="Concept 3", value= None, info="Pick concept type")
410
  with gr.Accordion("πŸ’‘ Or use a new concept πŸ‘‡", open=False):
411
  gr.Markdown("1. Upload a file with text variations of your concept (e.g. ask an LLM)")