Removed message about run button not working
Browse files
app.py
CHANGED
@@ -194,8 +194,8 @@ def run(path,progress_bar):
|
|
194 |
except Exception as e:
|
195 |
print(f"Error recloning repo: {e}")
|
196 |
if not fsave:
|
197 |
-
gr.Warning("Something went wrong with GitHub and data couldn't be sent to the server. Please check the logs
|
198 |
-
def download_data():
|
199 |
d = {'family':family,'alleles':alleles}
|
200 |
model_name = path
|
201 |
data_path = f'math/{model_name}.json'
|
@@ -208,7 +208,7 @@ def run(path,progress_bar):
|
|
208 |
json.dump(d,f)
|
209 |
# Provide the download link
|
210 |
return gr.File.update(value=path, label="Download data", file_name=f"{model_name}.json")
|
211 |
-
gr.Button("Download data",variant="primary").click(fn=download_data, inputs=[], outputs=None)
|
212 |
return None
|
213 |
|
214 |
|
@@ -373,7 +373,6 @@ with gr.Blocks(title="PhyloLM", theme=gr.themes.Default()) as demo:
|
|
373 |
"- If the model has already been processed, you'll be notified and no new run will start.\n"
|
374 |
"- If it hasn't been processed, it will be downloaded and be evaluated.\n\n"
|
375 |
"⚠️ Be careful when submitting large LLMs (typically >15B parameters) as they may exceed the GPU RAM or the time limit, leading to failed runs.\n"
|
376 |
-
" **UPDATE : The model sending module is under revision - it should be back again soon**"
|
377 |
)
|
378 |
|
379 |
with gr.Group():
|
|
|
194 |
except Exception as e:
|
195 |
print(f"Error recloning repo: {e}")
|
196 |
if not fsave:
|
197 |
+
gr.Warning("Something went wrong with GitHub and data couldn't be sent to the server. Please check the logs or report an issue. ")
|
198 |
+
"""def download_data():
|
199 |
d = {'family':family,'alleles':alleles}
|
200 |
model_name = path
|
201 |
data_path = f'math/{model_name}.json'
|
|
|
208 |
json.dump(d,f)
|
209 |
# Provide the download link
|
210 |
return gr.File.update(value=path, label="Download data", file_name=f"{model_name}.json")
|
211 |
+
gr.Button("Download data",variant="primary").click(fn=download_data, inputs=[], outputs=None)"""
|
212 |
return None
|
213 |
|
214 |
|
|
|
373 |
"- If the model has already been processed, you'll be notified and no new run will start.\n"
|
374 |
"- If it hasn't been processed, it will be downloaded and be evaluated.\n\n"
|
375 |
"⚠️ Be careful when submitting large LLMs (typically >15B parameters) as they may exceed the GPU RAM or the time limit, leading to failed runs.\n"
|
|
|
376 |
)
|
377 |
|
378 |
with gr.Group():
|