Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
@@ -5,6 +5,10 @@ import requests
|
|
5 |
import gradio as gr
|
6 |
|
7 |
STYLE = """
|
|
|
|
|
|
|
|
|
8 |
.group-border {
|
9 |
padding: 10px;
|
10 |
border-width: 1px;
|
@@ -184,7 +188,7 @@ def submit(
|
|
184 |
return f"something went wrong {response.status_code} = {response.text}"
|
185 |
|
186 |
with gr.Blocks(css=STYLE) as hf_endpoint:
|
187 |
-
with gr.Tab("π€ Inference Endpoint"):
|
188 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
189 |
|
190 |
with gr.Column(elem_classes=["group-border"]):
|
@@ -306,7 +310,7 @@ Name for your new endpoint""")
|
|
306 |
)
|
307 |
|
308 |
with gr.Column(elem_classes=["group-border"]):
|
309 |
-
with gr.Accordion("Serving Container", open=False):
|
310 |
with gr.Column():
|
311 |
gr.Markdown("""### Container Type
|
312 |
|
@@ -415,16 +419,16 @@ Name for your new endpoint""")
|
|
415 |
security_selector],
|
416 |
outputs=status_txt)
|
417 |
|
418 |
-
with gr.Tab("AWS"):
|
419 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
420 |
|
421 |
-
with gr.Tab("GCP"):
|
422 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
423 |
|
424 |
-
with gr.Tab("Azure"):
|
425 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
426 |
|
427 |
-
with gr.Tab("Lambdalabs"):
|
428 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
429 |
|
430 |
hf_endpoint.launch(enable_queue=True, debug=True)
|
|
|
5 |
import gradio as gr
|
6 |
|
7 |
STYLE = """
|
8 |
+
.no-border {
|
9 |
+
border: none;
|
10 |
+
}
|
11 |
+
|
12 |
.group-border {
|
13 |
padding: 10px;
|
14 |
border-width: 1px;
|
|
|
188 |
return f"something went wrong {response.status_code} = {response.text}"
|
189 |
|
190 |
with gr.Blocks(css=STYLE) as hf_endpoint:
|
191 |
+
with gr.Tab("π€ Inference Endpoint", elem_classes=["no-border"]):
|
192 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
193 |
|
194 |
with gr.Column(elem_classes=["group-border"]):
|
|
|
310 |
)
|
311 |
|
312 |
with gr.Column(elem_classes=["group-border"]):
|
313 |
+
with gr.Accordion("Serving Container", open=False, elem_classes=["no-border"]):
|
314 |
with gr.Column():
|
315 |
gr.Markdown("""### Container Type
|
316 |
|
|
|
419 |
security_selector],
|
420 |
outputs=status_txt)
|
421 |
|
422 |
+
with gr.Tab("AWS", elem_classes=["no-border"]):
|
423 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
424 |
|
425 |
+
with gr.Tab("GCP", elem_classes=["no-border"]):
|
426 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
427 |
|
428 |
+
with gr.Tab("Azure", elem_classes=["no-border"]):
|
429 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
430 |
|
431 |
+
with gr.Tab("Lambdalabs", elem_classes=["no-border"]):
|
432 |
gr.Markdown("# Deploy LLM on π€ Hugging Face Inference Endpoint", elem_classes=["center"])
|
433 |
|
434 |
hf_endpoint.launch(enable_queue=True, debug=True)
|