Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -14,7 +14,7 @@ else:
|
|
14 |
HF_TOKEN_ERROR = None
|
15 |
|
16 |
client = InferenceClient(token=HF_TOKEN)
|
17 |
-
PROMPT_IMPROVER_MODEL = "
|
18 |
|
19 |
def improve_prompt(original_prompt):
|
20 |
if HF_TOKEN_ERROR:
|
@@ -32,7 +32,7 @@ Improve this prompt: {original_prompt}
|
|
32 |
improved_prompt = client.text_generation(
|
33 |
prompt=prompt_for_llm,
|
34 |
model=PROMPT_IMPROVER_MODEL,
|
35 |
-
max_new_tokens=
|
36 |
temperature=0.7,
|
37 |
top_p=0.9,
|
38 |
repetition_penalty=1.2,
|
@@ -86,7 +86,7 @@ css = """
|
|
86 |
with gr.Blocks(css=css) as demo:
|
87 |
gr.Markdown(
|
88 |
"""
|
89 |
-
# Xylaria Iris
|
90 |
""",
|
91 |
elem_classes="title"
|
92 |
)
|
|
|
14 |
HF_TOKEN_ERROR = None
|
15 |
|
16 |
client = InferenceClient(token=HF_TOKEN)
|
17 |
+
PROMPT_IMPROVER_MODEL = "Qwen/Qwen2.5-Coder-32B-Instruct"
|
18 |
|
19 |
def improve_prompt(original_prompt):
|
20 |
if HF_TOKEN_ERROR:
|
|
|
32 |
improved_prompt = client.text_generation(
|
33 |
prompt=prompt_for_llm,
|
34 |
model=PROMPT_IMPROVER_MODEL,
|
35 |
+
max_new_tokens=1280,
|
36 |
temperature=0.7,
|
37 |
top_p=0.9,
|
38 |
repetition_penalty=1.2,
|
|
|
86 |
with gr.Blocks(css=css) as demo:
|
87 |
gr.Markdown(
|
88 |
"""
|
89 |
+
# Xylaria Iris v3
|
90 |
""",
|
91 |
elem_classes="title"
|
92 |
)
|