Update app.py
Browse files
app.py
CHANGED
@@ -14,8 +14,8 @@ model = AutoModelForCausalLM.from_pretrained('HuggingFaceM4/Florence-2-DocVQA',
|
|
14 |
processor = AutoProcessor.from_pretrained('HuggingFaceM4/Florence-2-DocVQA', trust_remote_code=True)
|
15 |
|
16 |
|
17 |
-
TITLE = "#
|
18 |
-
DESCRIPTION = "
|
19 |
|
20 |
|
21 |
colormap = ['blue','orange','green','purple','brown','pink','gray','olive','cyan','red',
|
@@ -74,9 +74,9 @@ with gr.Blocks(css=css) as demo:
|
|
74 |
gr.Examples(
|
75 |
examples=[
|
76 |
["hunt.jpg", 'What is this image?'],
|
77 |
-
["idefics2_architecture.png", 'How many tokens per image does it use?'],
|
78 |
-
["idefics2_architecture.png", "What type of encoder does the model use?"],
|
79 |
-
["image.jpg", "What's the share of Industry Switchers Gained?"]
|
80 |
],
|
81 |
inputs=[input_img, text_input],
|
82 |
outputs=[output_text],
|
|
|
14 |
processor = AutoProcessor.from_pretrained('HuggingFaceM4/Florence-2-DocVQA', trust_remote_code=True)
|
15 |
|
16 |
|
17 |
+
TITLE = "#Descripta"
|
18 |
+
DESCRIPTION = "Generate high-quality product descriptions from images and metadata"
|
19 |
|
20 |
|
21 |
colormap = ['blue','orange','green','purple','brown','pink','gray','olive','cyan','red',
|
|
|
74 |
gr.Examples(
|
75 |
examples=[
|
76 |
["hunt.jpg", 'What is this image?'],
|
77 |
+
# ["idefics2_architecture.png", 'How many tokens per image does it use?'],
|
78 |
+
# ["idefics2_architecture.png", "What type of encoder does the model use?"],
|
79 |
+
# ["image.jpg", "What's the share of Industry Switchers Gained?"]
|
80 |
],
|
81 |
inputs=[input_img, text_input],
|
82 |
outputs=[output_text],
|